Search Results for author: Segev Shlomov

Found 14 papers, 5 papers with code

We’ve had this conversation before: A Novel Approach to Measuring Dialog Similarity

no code implementations EMNLP 2021 Ofer Lavi, Ella Rabinovich, Segev Shlomov, David Boaz, Inbal Ronen, Ateret Anaby Tavor

The results demonstrate that our method outperforms the other approaches in capturing dialog flow, and is better aligned with the human perception of conversation similarity.

Towards Enterprise-Ready Computer Using Generalist Agent

no code implementations24 Feb 2025 Sami Marreed, Alon Oved, Avi Yaeli, Segev Shlomov, Ido Levy, Aviad Sela, Asaf Adi, Nir Mashkif

This paper presents our ongoing work toward developing an enterprise-ready Computer Using Generalist Agent (CUGA) system.

ST-WebAgentBench: A Benchmark for Evaluating Safety and Trustworthiness in Web Agents

1 code implementation9 Oct 2024 Ido Levy, Ben wiesel, Sami Marreed, Alon Oved, Avi Yaeli, Segev Shlomov

Autonomous web agents solve complex browsing tasks, yet existing benchmarks measure only whether an agent finishes a task, ignoring whether it does so safely or in a way enterprises can trust.

Autonomous Web Navigation

From Grounding to Planning: Benchmarking Bottlenecks in Web Agents

no code implementations3 Sep 2024 Segev Shlomov, Ben wiesel, Aviad Sela, Ido Levy, Liane Galanti, Roy Abitbol

We sharpen the distinction between the planning and grounding components and conduct a novel analysis by refining experiments on the Mind2Web dataset.

Benchmarking

SNAP: Semantic Stories for Next Activity Prediction

no code implementations28 Jan 2024 Alon Oved, Segev Shlomov, Sergey Zeltyn, Nir Mashkif, Avi Yaeli

To address this gap, we propose the novel SNAP method that leverages language foundation models by constructing semantic contextual stories from the process historical event logs and using them for the next activity prediction.

Activity Prediction Decision Making +2

Enhancing Trust in LLM-Based AI Automation Agents: New Considerations and Future Challenges

no code implementations10 Aug 2023 Sivan Schwartz, Avi Yaeli, Segev Shlomov

This paper explores these new challenges and opportunities, analyzes the main aspects of trust in AI agents discussed in existing literature, and identifies specific considerations and challenges relevant to this new generation of automation agents.

AI Agent

We've had this conversation before: A Novel Approach to Measuring Dialog Similarity

no code implementations12 Oct 2021 Ofer Lavi, Ella Rabinovich, Segev Shlomov, David Boaz, Inbal Ronen, Ateret Anaby-Tavor

The results demonstrate that our method outperforms the other approaches in capturing dialog flow, and is better aligned with the human perception of conversation similarity.

Phase Transitions in Kyle's Model with Market Maker Profit Incentives

no code implementations7 Mar 2021 Charles-Albert Lehalle, Eyal Neuman, Segev Shlomov

In addition to the classical framework, a revenue term is added to the market maker's performance function, which is proportional to the order flow and to the size of the bid-ask spread.

Not Enough Data? Deep Learning to the Rescue!

1 code implementation8 Nov 2019 Ateret Anaby-Tavor, Boaz Carmeli, Esther Goldbraich, Amir Kantor, George Kour, Segev Shlomov, Naama Tepper, Naama Zwerdling

Based on recent advances in natural language modeling and those in text generation capabilities, we propose a novel data augmentation method for text classification tasks.

Data Augmentation Deep Learning +7

Deep Dominance - How to Properly Compare Deep Neural Models

1 code implementation ACL 2019 Rotem Dror, Segev Shlomov, Roi Reichart

Comparing between Deep Neural Network (DNN) models based on their performance on unseen data is crucial for the progress of the NLP field.

The Hitchhiker's Guide to Testing Statistical Significance in Natural Language Processing

1 code implementation ACL 2018 Rotem Dror, Gili Baumer, Segev Shlomov, Roi Reichart

We establish the fundamental concepts of significance testing and discuss the specific aspects of NLP tasks, experimental setups and evaluation measures that affect the choice of significance tests in NLP research.

Survey

Cannot find the paper you are looking for? You can Submit a new open access paper.