{"@context":["https://w3id.org/fdo/context/v1",{"schema":"https://schema.org/","prov":"http://www.w3.org/ns/prov#","fdo":"https://w3id.org/fdo/vocabulary/"}],"@id":"https://fdo.portal.mardi4nfdi.de/fdo/Q5962720","@type":"DigitalObject","kernel":{"@id":"https://fdo.portal.mardi4nfdi.de/fdo/Q5962720","digitalObjectType":"https://schema.org/ScholarlyArticle","primaryIdentifier":"mardi:Q5962720","kernelVersion":"v1","immutable":true,"modified":"2025-05-08T16:23:45Z"},"profile":{"@context":"https://schema.org","@type":"ScholarlyArticle","@id":"https://portal.mardi4nfdi.de/entity/Q5962720","name":"Worst case complexity of direct search under convexity","headline":"Worst case complexity of direct search under convexity","description":"scientific article; zbMATH DE number 6544659","url":"https://portal.mardi4nfdi.de/entity/Q5962720","datePublished":"2016-02-23","author":[{"@id":"https://portal.mardi4nfdi.de/entity/Q245490"},{"@id":"https://portal.mardi4nfdi.de/entity/Q839811"}],"publisher":[{"@id":"https://portal.mardi4nfdi.de/entity/Q163006"}],"identifier":{"@type":"PropertyValue","propertyID":"doi","value":"10.1007/S10107-014-0847-0","url":"https://doi.org/10.1007/S10107-014-0847-0"},"sameAs":["https://doi.org/10.1007/S10107-014-0847-0"],"comment":"This paper considers directional direct-search methods applied to the unconstrained minimization of a real-valued, convex, and continuously differentiable objective function \\(f\\). It is proved that the direct-search methods of directional type, based on imposing sufficient decrease to accept new iterates, have the same worst case complexity bound and global rate of the gradient method for the unconstrained minimization of a convex and smooth function. The presented results are derived for convex functions where the supreme distance between any point in the initial level set and the solutions set is bounded. Such property is satisfied when the solutions set is bounded, but it is also met in several instances where the solutions sets are unbounded. It is shown that the number of iterations needed to reduce the norm of the gradient of the objective function below a certain threshold is at most proportional to the inverse of the threshold. It is also shown that the absolute error in the function values decay at a sublinear rate proportional to the inverse of the iteration counter. Finally, it is proved that the sequence of absolute errors of function values and iterates converges \\(r\\)-linearly in the strongly convex case.","citation":[{"@id":"https://portal.mardi4nfdi.de/entity/Q3083310"},{"@id":"https://portal.mardi4nfdi.de/entity/Q652287"},{"@id":"https://portal.mardi4nfdi.de/entity/Q2902870"},{"@id":"https://portal.mardi4nfdi.de/entity/Q3611732"},{"@id":"https://portal.mardi4nfdi.de/entity/Q4441973"},{"@id":"https://portal.mardi4nfdi.de/entity/Q2978646"},{"@id":"https://portal.mardi4nfdi.de/entity/Q2841063"},{"@id":"https://portal.mardi4nfdi.de/entity/Q5461068"},{"@id":"https://portal.mardi4nfdi.de/entity/Q3608990"},{"@id":"https://portal.mardi4nfdi.de/entity/Q4435425"},{"@id":"https://portal.mardi4nfdi.de/entity/Q1417731"},{"@id":"https://portal.mardi4nfdi.de/entity/Q2397749"},{"@id":"https://portal.mardi4nfdi.de/entity/Q2910875"},{"@id":"https://portal.mardi4nfdi.de/entity/Q2494519"},{"@id":"https://portal.mardi4nfdi.de/entity/Q5491447"},{"@id":"https://portal.mardi4nfdi.de/entity/Q5588250"},{"@id":"https://portal.mardi4nfdi.de/entity/Q743632"}]},"provenance":{"prov:generatedAtTime":"2025-05-08T16:23:45Z","prov:wasAttributedTo":"MaRDI Knowledge Graph"}}