scholarly journals Solving Moving-Blocks Problems

2017 ◽  
Author(s):  
André G. Pereira ◽  
Luciana S. Buriol ◽  
Marcus Ritt

Moving-blocks problems are extremely hard to solve and a representative abstraction of many applications. Despite their importance, the known computational complexity results are limited to few versions of these problems. In addition, there are no effective methods to optimally solve them. We address both of these issues. This thesis proves the PSPACE-completeness of many versions of moving-blocks problems. Moreover, we propose new methods to optimally solve these problems based on heuristic search with admissible heuristic functions and tie-breaking strategies. Our methods advance the state of the art, create new lines of research and improve the results of applications.

2013 ◽  
Vol 2013 ◽  
pp. 1-6
Author(s):  
Dunbo Cai ◽  
Sheng Xu ◽  
Tongzhou Zhao ◽  
Yanduo Zhang

Pruning techniques and heuristics are two keys to the heuristic search-based planning. Thehelpful actionspruning (HAP) strategy andrelaxed-plan-based heuristicsare two representatives among those methods and are still popular in the state-of-the-art planners. Here, we present new analyses on the properties of HAP. Specifically, we show new reasons for which HAP can cause incompleteness of a search procedure. We prove that, in general, HAP is incomplete for planning with conditional effects if factored expansions of actions are used. To preserve completeness, we propose a pruning strategy that is based onrelevance analysisandconfrontation. We will show that bothrelevance analysisandconfrontationare necessary. We call it theconfrontation and goal relevant actionspruning (CGRAP) strategy. However, CGRAP is computationally hard to be exactly computed. Therefore, we suggest practical approximations from the literature.


1968 ◽  
Vol 5 (04) ◽  
pp. 410-426
Author(s):  
Arthur Pitchersky ◽  
Arthur Southerland

The increasing demand for a flexible Naval response to a broad spectrum of military situations imposes a demand to carry out missions in increasingly higher sea states. Launching and retrieving buoyant objects or loading cargo into boats responding to the ocean-air interface requires improved technology for successful operations in high sea states. There is an urgent need for handling systems that provide the degree of control necessary for those Navy missions subjected to an increasing all-weather response. Advances in the state of the art or the development of new techniques are needed to support these operational requirements. This paper will discuss present handling systems and proposed new methods.


2010 ◽  
Vol 2 (1) ◽  
Author(s):  
Ludwig Zoeller

AbstractThis review paper intends to summarize the state of the art in loess research at the first international “Loess-fest’99” conference and to outline progress in loess research during the past decade. The focus is on loess as a terrestrial archive of climatic and environmental change during the Quaternary. The review highlights remarkable new results from regional investigations into European loess, as well as the emergence of new methods and refinements of established techniques, focussing on stratigraphy, dating and palaeoenvironment. It is concluded that loess research during the past decade not only has developed rapidly to take an outstanding place in Quaternary sciences, but also promises exciting perspectives for the next decade, in particular when combined approaches are applied to benefit from the now comprehensive pool of established and new methods.


Author(s):  
Daniel Höller ◽  
Pascal Bercher ◽  
Gregor Behnke ◽  
Susanne Biundo

Planning is the task of finding a sequence of actions that achieves the goal(s) of an agent. It is solved based on a model describing the environment and how to change it. There are several approaches to solve planning tasks, two of the most popular are classical planning and hierarchical planning. Solvers are often based on heuristic search, but especially regarding domain-independent heuristics, techniques in classical planning are more sophisticated. However, due to the different problem classes, it is difficult to use them in hierarchical planning. In this paper we describe how to use arbitrary classical heuristics in hierarchical planning and show that the resulting system outperforms the state of the art in hierarchical planning.


2016 ◽  
Vol 57 ◽  
pp. 229-271 ◽  
Author(s):  
Marcel Steinmetz ◽  
Jörg Hoffmann ◽  
Olivier Buffet

Unavoidable dead-ends are common in many probabilistic planning problems, e.g. when actions may fail or when operating under resource constraints. An important objective in such settings is MaxProb, determining the maximal probability with which the goal can be reached, and a policy achieving that probability. Yet algorithms for MaxProb probabilistic planning are severely underexplored, to the extent that there is scant evidence of what the empirical state of the art actually is. We close this gap with a comprehensive empirical analysis. We design and explore a large space of heuristic search algorithms, systematizing known algorithms and contributing several new algorithm variants. We consider MaxProb, as well as weaker objectives that we baptize AtLeastProb (requiring to achieve a given goal probabilty threshold) and ApproxProb (requiring to compute the maximum goal probability up to a given accuracy). We explore both the general case where there may be 0-reward cycles, and the practically relevant special case of acyclic planning, such as planning with a limited action-cost budget. We design suitable termination criteria, search algorithm variants, dead-end pruning methods using classical planning heuristics, and node selection strategies. We design a benchmark suite comprising more than 1000 instances adapted from the IPPC, resource-constrained planning, and simulated penetration testing. Our evaluation clarifies the state of the art, characterizes the behavior of a wide range of heuristic search algorithms, and demonstrates significant benefits of our new algorithm variants.


Sensors ◽  
2021 ◽  
Vol 21 (20) ◽  
pp. 6808
Author(s):  
Jianqiang Xiao ◽  
Dianbo Ma ◽  
Satoshi Yamane

Despite recent stereo matching algorithms achieving significant results on public benchmarks, the problem of requiring heavy computation remains unsolved. Most works focus on designing an architecture to reduce the computational complexity, while we take aim at optimizing 3D convolution kernels on the Pyramid Stereo Matching Network (PSMNet) for solving the problem. In this paper, we design a series of comparative experiments exploring the performance of well-known convolution kernels on PSMNet. Our model saves the computational complexity from 256.66G MAdd (Multiply-Add operations) to 69.03G MAdd (198.47G MAdd to 10.84G MAdd for only considering 3D convolutional neural networks) without losing accuracy. On Scene Flow and KITTI 2015 datasets, our model achieves results comparable to the state-of-the-art with a low computational cost.


Philosophies ◽  
2020 ◽  
Vol 5 (4) ◽  
pp. 37
Author(s):  
Attila Egri-Nagy ◽  
Antti Törmänen

The game of Go was the last great challenge for artificial intelligence in abstract board games. AlphaGo was the first system to reach supremacy, and subsequent implementations further improved the state of the art. As in chess, the fall of the human world champion did not lead to the end of the game. Now, we have renewed interest in the game due to new questions that emerged in this development. How far are we from perfect play? Can humans catch up? How compressible is Go knowledge? What is the computational complexity of a perfect player? How much energy is really needed to play the game optimally? Here, we investigate these and related questions with respect to the special properties of Go (meaningful draws and extreme combinatorial complexity). Since traditional board games have an important role in human culture, our analysis is relevant in a broader context. What happens in the game world could forecast our relationship with AI entities, their explainability, and usefulness.


2020 ◽  
Vol 62 (2) ◽  
pp. 99-115
Author(s):  
Janek Bevendorff ◽  
Tobias Wenzel ◽  
Martin Potthast ◽  
Matthias Hagen ◽  
Benno Stein

AbstractAuthorship verification is the task of determining whether two texts were written by the same author based on a writing style analysis. Author obfuscation is the adversarial task of preventing a successful verification by altering a text’s style so that it does not resemble that of its original author anymore. This paper introduces new algorithms for both tasks and reports on a comprehensive evaluation to ascertain the merits of the state of the art in authorship verification to withstand obfuscation.After introducing a new generalization of the well-known unmasking algorithm for short texts, thus completing our collection of state-of-the-art algorithms for verification, we introduce an approach that (1) models writing style difference as the Jensen-Shannon distance between the character n-gram distributions of texts, and (2) manipulates an author’s writing style in a sophisticated manner using heuristic search. For obfuscation, we explore the huge space of textual variants in order to find a paraphrased version of the to-be-obfuscated text that has a sufficiently high Jensen-Shannon distance at minimal costs in terms of text quality loss. We analyze, quantify, and illustrate the rationale of this approach, define paraphrasing operators, derive text length-invariant thresholds for termination, and develop an effective obfuscation framework. Our authorship obfuscation approach defeats the presented state-of-the-art verification approaches, while keeping text changes at a minimum. As a final contribution, we discuss and experimentally evaluate a reverse obfuscation attack against our obfuscation approach as well as possible remedies.


2013 ◽  
Vol 13 (4-5) ◽  
pp. 831-846 ◽  
Author(s):  
ESRA ERDEM ◽  
VOLKAN PATOGLU ◽  
ZEYNEP G. SARIBATUR ◽  
PETER SCHÜLLER ◽  
TANSEL URAS

AbstractWe study the problem of finding optimal plans for multiple teams of robots through a mediator, where each team is given a task to complete in its workspace on its own and where teams are allowed to transfer robots between each other, subject to the following constraints: 1) teams (and the mediator) do not know about each other's workspace or tasks (e.g., for privacy purposes); 2) every team can lend or borrow robots, but not both (e.g., transportation/calibration of robots between/for different workspaces is usually costly). We present a mathematical definition of this problem and analyze its computational complexity. We introduce a novel, logic-based method to solve this problem, utilizing action languages and answer set programming for representation, and the state-of-the-art ASP solvers for reasoning. We show the applicability and usefulness of our approach by experiments on various scenarios of responsive and energy-efficient cognitive factories.


2018 ◽  
Vol 2018 ◽  
pp. 1-11 ◽  
Author(s):  
Yu Sun ◽  
Rongrong Ni ◽  
Yao Zhao

In order to solve the problem of high computational complexity in block-based methods for copy-move forgery detection, we divide image into texture part and smooth part to deal with them separately. Keypoints are extracted and matched in texture regions. Instead of using all the overlapping blocks, we use nonoverlapping blocks as candidates in smooth regions. Clustering blocks with similar color into a group can be regarded as a preprocessing operation. To avoid mismatching due to misalignment, we update candidate blocks by registration before projecting them into hash space. In this way, we can reduce computational complexity and improve the accuracy of matching at the same time. Experimental results show that the proposed method achieves better performance via comparing with the state-of-the-art copy-move forgery detection algorithms and exhibits robustness against JPEG compression, rotation, and scaling.


Sign in / Sign up

Export Citation Format

Share Document