1,446
Views
2
CrossRef citations to date
0
Altmetric
Research Articles

An introduction to non-smooth convex analysis via multiplicative derivative

Pages 351-359 | Received 23 Oct 2018, Accepted 30 Jan 2019, Published online: 12 Feb 2019

Abstract

In this study, *-directional derivative and *-subgradient are defined using the multiplicative derivative, making a new contribution to non-Newtonian calculus for use in non-smooth analysis. As for directional derivative and subgradient, which are used in the non-smooth optimization theory, basic definitions and preliminary facts related to optimization theory are stated and proved, and the *-subgradient concept is illustrated by providing some examples, such as absolute value and exponential functions. In addition, necessary and sufficient optimality conditions are obtained for convex problems.

Mathematical Subject Classification (2010):

1. Introduction

Subdifferential, which is a very useful concept in non-smooth analysis and optimization theory, is a generalization of the classical differential. Some of the significant and widely used differentiability properties of subdifferential can be found in [Citation1–3]. Subdifferential is widely used for solving non-smooth optimization problems, some of them can be listed as follows [Citation3–15]. Of course, in addition to subgradient, there are several concepts, such as quasidifferential, discrete gradient and codifferential, and strategies, namely smoothing and scalarization techniques to develop optimization methods in the literature. In the literature, some well-known methods developed by using these concept and/or techniques are explained in [Citation16–34].

Conversely, the non-Newtonian calculus concept is based on the definition of multiplicative derivative. The multiplicative derivative is used to find the factor by which the value of a function changes as its variable changes. It is different than for the conventional derivative, which finds the rate at which the value of a function changes. For several real-life models such as population growth, growth and decay, and economic models wherein the dependent variables increase or decrease exponentially, it is more important to know the factor by which the variables change than the rate at which they change. Some real word motivation can be found in [Citation35–37]. In addition, these types of models are often encountered in non-smooth optimization problems.

To motivate the use of multiplicative derivative, we now present the following three applications. Two of them use discrete variables and the other use continuous variables. The first example involves the process of cell division, by which the number of cells in the body increases exponentially. First, a cell splits into two cells, which in turn split to form four cells, then eight and so on. Since the division process causes the number of cells to grow exponentially, multiplicative derivative is more appropriate than the conventional one for the continuous form of this model.

The second application is from the computational theory. Some computer algorithms require an exponentially increasing amount of resources (e.g., time, computer memory or number of function evaluations) although the problem size increases linearly. For instance, an algorithm for a problem of size n=1,2,3,,n may take 20,40,80,,2n10 seconds, respectively, for completion, thus making it difficult to solve the problem for more than 20 variables.

To give another example for the continuous case, assume that you deposit y0 $ in a bank account and it grows to y1 $ after 1 year. The annual growth rate is y1/y0, but what is the monthly growth rate? If the monthly rate is c, then after 12 months the total amount will be y1=y0c12. In other words, the monthly rate c is (y1/y0)1/12. If we instead assume that the balance is updated daily or hourly, we obtain c=(y1/y0)1/365 or c=(y1/y0)1/8760, respectively. By expressing the balance over time as the function f, we obtain the formula c=(f(x+h)/f(x))1/h for appropriate values of h. In other words, if we are working in terms of months, then c=(f(12)/f(0))1/12=(y1/y0)1/12. Moreover, if we are working in terms of days, then c=(f(365)/f(0))1/365=(y1/y0)1/365. Finally, if the balance is updated continuously, then the rate of change is c=(f(x+h)/f(x))1/h, in the limit as 1/h tends to zero.

Due to aforementioned three applications and their reasons, Michael Grossman and Robert Katz [Citation38] defined and used *-differential to construct a non-Newtonian calculus in 1972. Later, another brief paper [Citation39] was published in 1999. Since then, several applications of the *-derivative have been presented [Citation40–44].

Our main aim in this paper is to generalize multiplicative derivative to convex functions and state the properties of the derivative. This generalization contributes to the development of non-Newtonian calculus and can be used in the non-smooth optimization theory.

The paper is organized as follows. In Section 2, the multiplicative derivative and *-gradient are recalled. After that, *-directional derivative and *-subdifferential are defined in Section 3. Then, their properties are stated, proved and discussed. By using this new concept, a optimality condition for convex optimization problem is given in Section 4. Section 5 concludes the paper.

2. Preliminaries

In this section, we provide some basic information about the multiplicative derivative, i.e. non-Newtonian calculus, which can be found in [Citation38,Citation39,Citation41].

Definition 2.1

Assume that the function f:IRIR is positive valued. If the limit limh0f(x+h)f(x)1/h exists, then f is said to be *-differentiable at the point x. The value of this limit is known as the *-derivative (or multiplicative derivative) of the function f at the point x and is denoted by f(x) or df(x)/dx.

Remark 2.1

Since f is a positive-valued function, it can easily be seen from Definition 2.1 that f(x)0.

By using Definition 2.1, *-derivative can be extended to functions of several variables. Let us consider the function f:IRnIR of n variables. The partial *-derivative of f with respect to xi for i{1,2,,n} can then be defined by fixing all other variables xk, where k{1,2,,i1,i+1,,n} and is denoted by fxi, fi or f/xi.

Definition 2.2

Assume that the function f:IRnIR is positive valued. If all the partial *-derivatives of f exist, then the *-gradient of f is a vector-valued function and is defined as fx1(x),fx2(x),,fxn(x) and is denoted by f(x).

The relationship between the *-derivative and the classical derivative was given in [Citation39] as (1) df(x)dx=ef(x)/f(x).(1) In [Citation41], the multiplicative chain rule for functions of two variables is given without proof. This rule can be generalized to functions of n variables, as shown in the following theorem.

Theorem 2.1

Let f:IRnIR be a positive-valued function with continuous partial *-derivatives and xi:IRIR be differentiable functions for i{1,,n}. Then, df(x1(t),,xn(t))dt=fx1(x1(t),,xn(t))x1(t)fxn(x1(t),,xn(t))xn(t).

Proof.

From Equation (Equation1), we obtain that df(x)dt:=df(x1(t),,xn(t))dt=edf(x1(t),,xn(t))d(t)f(x1(t),,xn(t)). By using the classical chain rule, we then find df(x)dt=ef(x1(t),,xn(t))x1dx1(t)dt++f(x1(t),,xn(t))xndxn(t)dtf(x1(t),,xn(t))=ef(x1(t),,xn(t))x1dx1(t)dtf(x1(t),,xn(t))ef(x1(t),,xn(t))xndxn(t)dtf(x1(t),,xn(t))=fx1(x1(t),,xn(t))x1(t)fxn(x1(t),,xn(t))xn(t).

3. *-Directional derivative and *-subdifferential

The partial *-derivative shows how many times the value of a function changes along one of its coordinates while all other coordinates are constant. Since in optimization theory, this change needs to be directional in any direction v, the following definition can be given.

Definition 3.1

Let f:IRnIR be a positive valued function, and vIRn be a non-zero vector. The function f is said to be *-directionally differentiable at a point x in the direction v, if there exists the finite limit f(x;v)=limh0f(x+hv)f(x)1/h, which is called *-directional derivative of f at x in the direction v.

Then, using the following theorem, we can find the relation between *-directional derivative and directional derivative.

Theorem 3.1

Let f:IRnIR be a positive function. If f is directionally differentiable at a point x in a direction v, then f is *-directionally differentiable at x in the direction v, and f(x;v)=ef(x;v)f(x), where f(x;v) denotes the directional derivative of the function f at the point x in the direction v.

Proof.

By the definition of *-directional derivative, f(x;v)=limh0f(x+hv)f(x)1h=limh0f(x+hv)f(x)f(x)+1f(x+hv)f(x)f(x)f(x)f(x+hv)f(x)f(x+hv)f(x)h1f(x). Saying t=f(x)/f(x+hv)f(x), one can show that easily t as h0. Since f is directionally differentiable at the point x in the direction v, limh0f(x+hv)f(x)/h exist and equals f(x;v). Thus f(x;v)=limt1+1tf(x;v)t/f(x)=ef(x;v)f(x) ,since limt1+1tt=e.

Let us define (2) uv=u1v1u2v2unvn(2) for u=(u1,,un)IR+n and v=(v1,,vn)IRn, where IR+n={x=(x1,x2,,xn)IRn|xi0 i{1,2,,n}}. By using this notation, we can give the connection between *-directional derivative f(x;v) and *-gradient f(x) in the following theorem.

Theorem 3.2

Let f:IRnIR be a positive *-differentiable function. Then f(x;v)=f(x)v.

Proof.

Similarly, the proof of Theorem 2.1, the proof of this theorem can be obtained from Theorem 3.1.

For the convex functions, the *-directional derivative has the following features.

Theorem 3.3

If f:IRnIR is a positive convex function, then the *-directional derivative f(x;v) exists in every direction vIRn and defined as f(x;v)=infh>0e1hf(x+hv)f(x)1.

Proof.

Let us define the function ψ:IRIR for an arbitrary vector vIRn, as ψ(h)=1hf(x+hv)f(x)f(x). Let ϵ>0, without loss of generality we can choose h1,h2(0,ϵ) such that h1<h2. Then, ψ(h2)ψ(h1)=1h1h2h1f(x+h2v)h1f(x)f(x)h2f(x+h1v)h2f(x)f(x)=1h1h1h2f(x+h2v)1h1h2f(x)f(x+h1v)f(x). Since h1<h2, h1/h2<1. Then, by using convexity of the function f(x), it can be easily seen that h1h2f(x+h2v)+1h1h2f(x)fh1h2(x+h2v)+1h1h2x=f(x+h1v). Thus ψ(h2)ψ(h1)1h1f(x+h1v)f(x+h1v)f(x)=0. This means that function ψ decreases as h0.

On the other hand, since we study in the case h0, we can assume 0<h<ϵ<1. Let us show that ψ(h) is bounded below for h(0,ϵ)(0,1). For this purpose, for 0<h<ϵ observe that ψ(h)ψ(ϵ/2)=1/hf(x+hv)f(x)f(x)2ϵf(xϵ2v)f(x)f(x)=f(x+hv)f(x)+2hϵf(xϵ2v)2hϵf(x)hf(x)=212f(x+hv)+12f(x)+hϵf(xϵ2v)+(1hϵ)f(x)2f(x)hf(x). Since f is convex, ψ(h)ψ(ϵ/2)2f(x+h2v)+f(xh2v)2f(x)hf(x)=412f(x+h2v)+12f(xh2v)f(x)hf(x)4f(x)f(x)hf(x)0. Since the above inequality satisfied for all ε, the function ψ is bounded below. Therefore since ψ is decreasing as h0 and bounded below for h(0,ϵ)(0,1), limh0ψ(h) exists and (3) limh0ψ(h)=infh>0ψ(h).(3) From the definition of *-derivative, one has f(x;v)=limh0f(x+hv)f(x)1/h=limh0f(x+hv)f(x)f(x)+1f(x)f(x+hv)f(x)ψ(h). Since limh0ψ(h) exists and limh0((f(x+hv)f(x)/f(x))+1)f(x)/f(x+hv)f(x)=e, we can conclude that f(x;v)=limh0eψ(h). From (Equation3), we obtain f(x;v)=infh>0eψ(h). Hence f(x;v)=infh>0e1hf(x+hv)f(x)1.

Remark 3.1

As it is seen from the proof of Theorem 3.3, for any ϵ>0 one has f(x;v)=infh(0,ϵ)e1hf(x+hv)f(x)1.

In the following theorems, we will show that *-directional derivative satisfies some good properties which are necessary for calculation and most of them obtained from Theorem 3.3. When using Theorem 3.3, it is assumed that h(0,1) according to Remark 3.1.

Theorem 3.4

Let f:IRnIR be a positive function. Then for the function vf(x;v), the following equation f(x;λv)=f(x;v)λfor0<λIR holds.

Proof.

From Definition 3.1, f(x;λv)=limh0f(x+hλv)f(x)1/h=limh0f(x+hλv)f(x)1/λhλ. Let k:=λh. Since 0<λ and h0, then k0. Thus f(x;λv)=limk0f(x+kv)f(x)1/kλ=f(x;v)λ.

Theorem 3.5

Let f:IRnIR be a positive convex function. Then the function vf(x;v) satisfies the following properties:

  1. f(x;v+w)f(x;v)f(x;w),

  2. f(x;v)f(x;v)1.

Proof.

(a) By using Theorem 3.3, we obtain f(x;v+w)=infh>0e1hf(x+h(v+w))f(x)1=infh>0e1hf12(x+2hv)+12(x+2hw)f(x)1. Since f is convex and h<1, we get f(x;v+w)infh>0e1h12f(x+2hv)+12f(x+2hw)f(x)1=infh>0e12hf(x+2hv)f(x)1+f(x+2hw)f(x)1=infh>0e12hf(x+2hv)f(x)1e12hf(x+2hw)f(x)1infh>0e12hf(x+2hv)f(x)1infh>0e12hf(x+2hw)f(x)1=f(x;v)f(x;w). (b) From the part (a), we have f(x;vv)f(x;v)f(x;v). As f(x;0)=1, we have f(x;v)f(x;v)1.

Since a convex function f:IRnIR is locally Lipschitz continuous at any point xIRn (for proof see [Citation1]), the following properties of *-directional derivative can be presented.

Theorem 3.6

Let f:IRnIR be a positive valued convex function with a Lipschitz constant K at xIRn. Then the function vf(x;v) satisfies the following inequality: lnf(x;v)Kvf(x).

Proof.

From Theorem 3.3, one can see that f(x;v)=infh>0e1hf(x+hv)f(x)1=infh>0e1hf(x+hv)f(x)f(x)infh>0e1hKx+hvxf(x)infh>0eKvf(x)=eKvf(x) or lnf(x;v)Kvf(x).

In the literature, the subgradient and subdifferential of a convex function are defined from the fact that f(y)f(x)+f(x)T(yx) under the differentiability assumption. Clarke [Citation2] defined the subgradient and the subdifferential by using this inequality. Now, we are going to define the *-subgradient and *-subdifferential, but before that we need to mention a similar inequality which hold for the *-gradient.

Theorem 3.7

Let f:IRnIR be a positive valued convex function. If f has all its partial *-derivatives, then the following inequality: (4) yIRnf(y)f(x)+lnf(x)f(x)(yx)(4) is satisfied.

Proof.

It is clear that if y=x, then inequality (Equation4) holds. Thus, assume that xy. By Theorem 3.3, we can write f(x)yx=f(x;yx). Then, using Theorem 3.3, it can be written as f(x)yx=infh>0e1hf(x+h(yx))f(x)1=infh>0e1hf((1h)x+hy)f(x)1. Since f is convex and h<1, we have f(x)yxinfh>0e1h((1h)f(x)+hf(y)f(x))1=ef(x)+f(y)f(x) or f(y)f(x)+lnf(x)f(x)(yx).

Remark 3.2

The term lnf(x)f(x)(yx) in the right-hand side of inequality (Equation4) in Theorem 3.7 is not equal to f(x)(yx)lnf(x), however, lnf(x)f(x)(yx)=f(x)lnf(x)(yx). The reason of this difference is that (yx)IRn while f(x)IR and moreover f(x)f(x)(yx) is real number which is defined as in (Equation2).

Definition 3.2

The *-subdifferential of a convex function f:IRnIR at the point xIRn is the set f(x) of vectors vIR+n such that (5) f(x)=vIR+nf(y)f(x)+lnvf(x)(yx)for allyIRn,(5) where the vector vf(x) is called *-subgradient.

Now, some examples can be given to visualize the *-subdifferential concept.

Example 3.1

Let us consider the function f(x)=|x|+1. Even if the function f is continuous, it is not *-differentiable at x=0. Indeed, limh0f(h)f(0)1/h=limh0h+111/hlimh0h+111/h=limh0h+11/h=e1,limh0+h+111/h=limh0+h+11/h=e. Thus the function f is not *-differentiable at x=0. On the other hand, since the function f is convex, *-subdifferential of the function f at the point x=0 exists and f(0)=vIR+y+11+lnvyfor all yIR=vIR+yylnvfor allyIR=e1,e. Now, let us find the *-subdifferential of the function f at the point x where the function f is *-differentiable. Using Theorem 3.2, we can write *-subdifferential of f at the point x>0 as f(x)=limh0f(x+h)f(x)1/h=limh0x+h+1x+11/h=e1/x+1. Then, let us calculate the *-subdifferential f(x) for x>0, if vf(x), f(y)f(x)+lnvf(x)(yx)for allyIRy+1x+1+lnvx+1(yx)for allyIRyxx+1(yx)lnvfor allyIRyxx+1(yx)lnvfor ally0,yxx+1(yx)lnvfor ally<0. By solving the first inequality for two cases, namely, y>x and y<x, we find v=e1/x+1. Then if we substitute v=e1/x+1 in the second inequality, we obtain yy which is always true for y<0. Thus f(x)={e1/x+1}.

Hence the result f(x)={f(x)} obtained from Example 3.1 can be generalized for functions which are *-differentiable at the point x. Consequently, the following theorem can be given.

Theorem 3.8

If a positive convex function f is *-differentiable at the point x, then the *-subdifferential f(x) is a singleton set consisting of f(x), i.e. f(x)={f(x)}.

Before the proof of this theorem, we need to give the following lemma.

Lemma 3.1

Let f:IRnIR be a positive convex function. Then for all xIRn, f(x)={vIRn|f(x;d)vdfor all dIRn}.

Proof.

Let S:={vIR+n|f(x;d)vdfor all dIRn} and vS. Then we have vdf(x;d). From Theorem 3.3, one can obtain that vd=infh>0e1hf(x+hd)f(x)1=infh>0e1hf((1h)x+h(x+d)f(x)1. Convexity of f yields that vdinfh>0e1h(1h)f(x)+hf(x+d)f(x)1=infh>0ef(x+d)f(x)f(x)=ef(x+d)f(x)f(x). Therefore we get f(x+d)f(x)+f(x)lnvd=f(x)+lnvf(x)d, where h1. If we choose d:=yx, it can be clearly seen that vf(x).

Conversely, if vf(x), then it follows from Theorem 3.3 that f(x;d)=infh>0e1hf(x+hd)f(x)1infh>0e1hlnvhd,sincevf(x)=vd. Thus vS.

Now, we can give the proof of Theorem 3.8.

Proof.

From Theorem 3.3, the *-directional derivative f(x;d) of a convex function f exists in every direction dIRn. By Theorem 3.2, since f(x;d)=(f(x))d for all dIRn, from Lemma 3.1, f(x)f(x). Now, if we assume that there exists vf(x) such that vf(x), then by using Lemma 3.1 we have vdf(x;d)=f(x)d. If we choose d=(1,0,,0) and d=(1,0,,0), we obtain v1fx1(x) and v1fx1(x), respectively, which means v1=fx1(x). Similarly, for d=(0,1,0,,0) and d=(0,1,0,,0), we obtain v2=fx2(x). Consequently, for all i{1,2,,n}, we get vi=fxi(x), that is, v=f(x). However, this contradicts the assumption that vf(x). Therefore f(x)={f(x)}.

Example 3.2

In this example, we consider function f(x)=e|x| and try to find the *-subdifferential of this function. Since the function f is continuous except x=0, by Theorem 3.8, f(x)={f(x)}={e},x>0,{e1},x<0.

Indeed, f(x)=limh0f(x+h)f(x)1/h=limh0ex+hex1/h. This implies that if x>0, then f(x)=e and if x<0, then f(x)=e1. Now, we compute f(0). If vf(0), then yIR f(y)f(0)+lnvf(0)(y0)ey1lnvyeey1yv, if y>0,eey1yv,if y<0,e1ve, where we have used infy>0{eey1y}=e and supy<0{eey1y}=e1. Consequently, f(x)={e1},x<0,e1,e,x=0,{e},x>0.

As it is seen in this example, it is not an easy task to compute *-subgradient at a non-differential point. This is an expected difficulty which is caused by the nature of non-smoothness and the subgradient itself already has the same difficulty. Alternatively *-subgradient can be used to determine whether a given point is an extremum point or not, which is discussed in Section 4. Although subgradient can be used to determine being an extremum point, *-subgradient is also useful for exponential functions since the rate of growth increases rapidly. On the other hand, the use of *-subgradient in optimization algorithms for some class of optimization problems may give us more efficient results. At this point, it may be useful to give the following theorem putting the relation between subgradient and *-subgradient.

Theorem 3.9

Let f:IRnIR be a positive valued convex function. Then v=v1, v2,, vnf(x)ev1/f(x), ev2/f(x),, evn/f(x)f(x).

Proof.

First of all, let's recall the definition of the subgradient f(x) for a convex function f:IRnIR which is f(x)=vIRnf(y)f(x)+vT(yx)for allyIRn. Assume vf(x). We should show that the vector (ev1/f(x),ev2/f(x),, evn/f(x)) satisfies the inequality in Definition 3.2. Since vf(x) where x=(x1,x2,,xn), for all y=(y1,y2,,yn)IRn, we have f(y)f(x)+vT(yx)=f(x)+(v1(y1x1)+v2(y2x2)++vn(xnyn))=f(x)+lnev1(y1x1)+v2(y2x2)++vn(xnyn)=f(x)+lnev1(y1x1)ev2(y2x2)evn(xnyn)=f(x)+lnev1/f(x), ev2/f(x),, evn/f(x)f(x)(yx). Therefore, (ev1/f(x), ev2/f(x),, evn/f(x))f(x.) Similarly, other direction of Theorem Equation1 can be shown.

As a result of this theorem, one can say that f(x) is a non-empty compact set, since the subdifferential f(x) of a convex function f(x) is a non-empty compact set [Citation1]. Being a non-empty set is obvious. For compactness of the set f(x), recall that any continuous image of a compact set is compact. This fact can be given as a following corollary.

Corollary 3.1

The set f(x) is a non-empty compact set in IRn.

4. Optimality condition for convex problem

Let us give a optimality condition via *-subgradient for the following non-smooth unconstrained optimization problem (6) minimizef(x),subject toxIRn,(6) where f:IRnIR is a convex function. This problem is also known as a convex problem.

Definition 4.1

A point xIRn is a global optimum of the problem (Equation6) if xIRn f(x)f(x) is satisfied.

In general, the following definition helps us to find the point satisfying the necessary optimality condition. For convex case, it gives us an idea whether the point is global optimum or not.

Definition 4.2

A point xIRn satisfying 1nf(x) is called a *-stationary point of f. Here, 1n denotes the point in IRn such that all entries are 1, i.e. 1n=(1,1,,1)IRn.

The following theorem presents the relation between global optimum and *-stationary point for the unconstrained non-smooth convex problem (Equation6).

Theorem 4.1

x is a global solution of the non-smooth optimization problem (Equation6) if and only if x is a *-stationary point of f.

Proof.

First, assume that x is a *-stationary point of f. Thus 1nf(x). Then for all xIRn, f(x)f(x)+ln1nf(x)(xx), which implies that for all xIRn, f(x)f(x) since ln1nf(x)(xx)=0. Consequently, x is a global solution of the non-smooth optimization problem (Equation6).

Conversely, assume that x is a global solution of the non-smooth optimization problem (Equation6). Then, f(x+hd)f(x) for all dIRn and hIR, which implies (f(x+hd)/f(x))10. From Theorem 3.3, f(x;d)=infh>0e1hf(x+hd)f(x)1infh>0e0=1. On the other hand, from Lemma 3.1, vf(x) if and only if f(x;d)vd for all dIRn. Since f(x;d)1, then vd=1 for all dIRn. Consequently, v=1n=(1,1,,1)f(x).

Theorem 4.1 tells us that the *-subgradient can be used to develop new optimization algorithms.

5. Conclusion

In this study, a new concept known as *-subgradient has been introduced with the help of the multiplicative derivative for use in the non-smooth optimization theory. To achieve this, *-directional derivative has been defined for positive-valued functions, and its properties have been proven. Since optimization theory deals with finding minimum function values, the objective function should be bounded below; otherwise, the function will not have a global optimum. In light of this fact, one can make the function positive by adding a sufficiently large number M so that f(x)+M0. Then, an inequality involving the *-gradient was stated and proved. Using this inequality, we were then able to define *-subgradient similar to the way subgradient has been defined in the literature. In addition, the relationship between *-directional derivative and *-subgradient has been presented. Finally, an optimality condition for non-smooth unconstrained convex optimization problems has been proven.

In summary, *-subgradient can be used to develop new methods in optimization theory and can be generalized to local Lipschitz continuous functions in an analogous way. It may have some advantage for some type of optimization problems, for example, if you use *-subgradient for any method, you study on large values instead of on small values. Thus it is good idea to develop new methods for functions whose values are small.

Acknowledgments

I would like to thank anonymous referees for their criticism and comments which helped the authors to significantly improve the quality of the paper.

Disclosure statement

No potential conflict of interest was reported by the authors.

References

  • Bagirov A, Karmitsa N, Mäkelä MM. Introduction to nonsmooth optimization. Cham/Heidelberg/New York/Dordrecht London: Springer; 2014. (Theory, Practice and Software).
  • Clarke FH. Optimization and nonsmooth analysis. New York: John Wiley & Sons Inc.; 1983. (Canadian Mathematical Society Series of Monographs and Advanced Texts). A Wiley-Interscience Publication.
  • Mäkelä MM, Neittaanm”aki P. Nonsmooth optimization. River Edge, NJ: World Scientific Publishing Co. Inc.; 1992. (Analysis and algorithms with applications to optimal control).
  • Anstreicher KM, Wolsey LA. Two ‘well-known’ properties of subgradient optimization. Math Program. 2009;120(1, Ser. B):213–220.
  • Beltran C, Heredia FJ. An effective line search for the subgradient method. J Optim Theory Appl. 2005;125(1):1–18.
  • Bertsekas DP. Nonlinear programming. 2nd ed. Belmont (MA): Athena Scientific; 1999.
  • Gaudioso M, Monaco MF. A bundle type approach to the unconstrained minimization of convex nonsmooth functions. Math Programming. 1982;23(2):216–226.
  • Hiriart-Urruty JB, Lemarechal C. Convex analysis and minimization algorithms. Vols. 1 and 2, Heidelberg: Springer-Verlag; 1993.
  • Kiwiel KC. Methods of descent for nondifferentiable optimization. Berlin, Springer-Verlag; 1985. (Lecture Notes in Mathematics; 1133)
  • Mifflin R. An algorithm for constrained optimization with semismooth functions. Math Oper Res. 1977;2(2):191–207.
  • Nedić A, Ozdaglar A. Subgradient methods for saddle-point problems. J Optim Theory Appl. 2009;142(1):205–228.
  • Nesterov Y. Primal-dual subgradient methods for convex problems. Math Program. 2009;120(1, Ser. B):221–259.
  • Shor NZ. Minimization methods for nondifferentiable functions. Berlin: Springer-Verlag; 1985. (Springer Series in Computational Mathematics; 3). Translated from the Russian by K. C. Kiwiel and A. Ruszczyński.
  • Shor NZ. Nondifferentiable optimization and polynomial problems. Dordrecht: Kluwer Academic Publishers; 1998. (Nonconvex Optimization and its Applications; 24)
  • Wolfe P. A method of conjugate subgradients for minimizing nondifferentiable functions. Math Program Stud. 1975;3:145–173. Nondifferentiable optimization
  • Akteke-Ozturk B, Koksal G, Weber GW. Nonconvex optimization of desirability functions. Qual Eng. 2018;30(2):293–310.
  • Akteke-Öztürk B, Weber G-W, Köksal G.. Optimization of generalized desirability functions under model uncertainty. Optimization. 2017;66(12):2157–2169.
  • Akteke-Ozturk B, Weber G-W, Kropat E. Continuous optimization approaches for clustering via minimum sum of squares. In: Proc. 20th Mini-EURO conf. continuous optimization and knowledge-based technologies; 2007.
  • Bagirov A, Ugon J. Codifferential method for minimizing nonsmooth dc functions. J Glob Optim. 2011;50(1):3–22.
  • Bagirov AM. Numerical methods for minimizing quasidifferentiable functions: a survey and comparison. In: Quasidifferentiability and related topics. Dordrecht: Kluwer Acad. Publ.; 2000. (Nonconvex Optim. Appl. Vo. 43), p. 33–71.
  • Bagirov AM, Ganjehlou AN. A quasisecant method for minimizing nonsmooth functions. Optim Methods Softw. 2010;25(1):3–18.
  • Bagirov AM, Ganjehlou AN, Tor H, Ugon J. A generalized subgradient method with piecewise linear subproblem. Dyn Contin Discrete Impuls Syst Ser B Appl Algorithms. 2010;17(5):621–638.
  • Bagirov AM, Ghosh M, Webb D. A derivative-free method for linearly constrained nonsmooth optimization. J Ind Manag Optim. 2006;2(3):319–338.
  • Bagirov AM, Karas”ozen B, Sezer M. Discrete gradient method: derivative-free method for nonsmooth optimization. J Optim Theory Appl. 2008;137(2):317–334.
  • Ben-Tal A, Teboulle M. A smoothing technique for nondifferentiable optimization problems. In: Optimization (Varetz, 1988). Berlin: Springer; 1989. (Lecture Notes in Math.; 1405), p. 1–11.
  • Conn AR, Toint P. An algorithm using quadratic interpolation for unconstrained derivative free optimization. In G. Di Pillo and F. Giannessi, editors, Nonlinear optimization and applications. New York: Kluwer Academic/Plenum Publishers; 1996. p. 27–47.
  • Demyanov VF, Bagirov AM, Rubinov AM. A method of truncated codifferential with application to some problems of cluster analysis. J Global Optim. 2002;23(1):63–80.
  • Dem'yanov VF, Dolgopolik MV. Codifferentiable functions in Banach spaces: methods and applications to problems of variation calculus.. Vestnik Sankt-Peterburgskogo Universiteta. Seriya 10. Prikladnaya Matematika. Informatika. Protsessy Upravleniya. 2013;3:48–66.
  • Dolgopolik MV. Abstract convex approximations of nonsmooth functions. Optimization. 2014;1–31.
  • Dolgopolik MV. Inhomogeneous convex approximations of nonsmooth functions. Russ Math. 2012;56(12):28–42.
  • Koksoy O. Dual response optimization: the desirability approach. Int J Ind Eng Theory, Appl Pract. 2005;12(4):335–342.
  • Miettinen K. Nonlinear multiobjective optimization, Vol. 12. Springer Science & Business Media, 2012.
  • Nesterov Y. Smooth minimization of non-smooth functions. Math Program. 2005;103(1, Ser. A):127–152.
  • Polak E, Royset JO, Womersley RS. Algorithms with adaptive smoothing for finite minimax problems. J Optim Theory Appl. 2003;119:459–484. doi:10.1023/B:JOTA.0000006685.60019.3e.
  • FoxJr WW. An exponential surplus-yield model for optimizing exploited fish populations. Trans Am Fish Soc. 1970;99(1):80–88.
  • Lippman SA. Applying a new device in the optimization of exponential queuing systems. Oper Res. 1975;23(4):687–710.
  • Ross SM. Applied probability models with optimization applications. New York: Courier Corporation; 2013.
  • Grossman M, Katz R. Non-Newtonian calculus: A self-contained, elementary exposition of the authors' investigations. Pigeon Cove (MA): Lee Press; 1972.
  • Stanley D. A multiplicative calculus. Probl Res Issues Math Undergrad Stud. 1999;9(4):310–326.
  • Aniszewska D. Multiplicative Runge–Kutta methods. Nonlinear Dyn. 2007;50(1):265–272.
  • Bashirov AE, Kurpınar EM, Özyapıcı A. Multiplicative calculus and its applications. J Math Anal Appl. 2008;337(1):36–48.
  • Misirli E, Ozyapici A. Exponential approximations on multiplicative calculus. Proc Jangjeon Math Soc. 2009;12(2):227–236.
  • Ozyapici A, Bilgehan B. Finite product representation via multiplicative calculus and its applications to exponential signal processing. Numer Algorith. 2016;71(2):475–489.
  • Riza M, Akt”ore H. The Runge–Kutta method in geometric multiplicative calculus. LMS J Comput Math. 2015;18(1):539–554.