Heads up for a hot new root seeking algorithm!!

01232017, 11:11 AM
(This post was last modified: 01232017 11:12 AM by Dieter.)
Post: #21




RE: Heads up for a hot new root seeking algorithm!!
(01212017 02:35 PM)Namir Wrote: Regarding the link to the article you mentioned in your message. Can you check the new algorithm by the author Thukral. I implemented equation 7 in the article and got bizarre results! Perhaps I am doing something wrong? When I replaced the second subtraction in equation 7 with a multiplication, the algorithm worked but was painfully slow to converge! Ad fontes! Always use the original source, i.e. the paper mentioned in footnote 7. Indeed there is an error in equation (20) of the quoted text. The correct formula has the derivative \(f'(x_n)^2\) as the first term of the first nominator, and not the function \(f(x_n)^2\) as shown in the linked article. Dieter 

01232017, 02:22 PM
Post: #22




RE: Heads up for a hot new root seeking algorithm!!
(01212017 02:35 PM)Namir Wrote: I suspect typos om the article since the title has one " Nonlinear Equations of Type f(0)=0" instead of " Nonlinear Equations of Type f(x)=0". Such Thukral algorithm is more or less useless, as it only works for zero roots. It states in the abstract that it can be used solely when the root is zero. Thus, the title f(0) = 0 is correct, the algorithm only converges to a root (yeah, with order 3 and only one function evaluation per iteration) if such root is zero. I posted it just for the tables, not for the algorithm. I'm working now on some Python code to test different algorithms. Up to now I was able to check that the Newton algorithm has indeed a 2nd order of convergence (I checked it working with 16, 32, 64, 128, 256, 512 & 1024 decimal digits). I'll test it with more equations and also will check the Halley & Ostrowsky methods. If all goes right, I can check your algorithm. Regards. César  Information must flow. 

01232017, 06:23 PM
Post: #23




RE: Heads up for a hot new root seeking algorithm!!
(01232017 02:22 PM)emece67 Wrote: Such Thukral algorithm is more or less useless, as it only works for zero roots. It states in the abstract that it can be used solely when the root is zero. Thus, the title f(0) = 0 is correct, the algorithm only converges to a root (yeah, with order 3 and only one function evaluation per iteration) if such root is zero. We are talking about two different methods here. The paper on sapub.org shows a current (2016) paper by R. Thukral, quoting an older paper of the same author from 2008. Here equation [20] is the 2008 Thukral formula (with an error). This formula works for nonzero roots. Because this older Thukral formula [20] has an error Namir does not get the expected results. After a correction (cf. the original source, i.e. the 2008 Thukral paper) the formula works fine and performs similar to the other posted thirdorder methods, sometimes even better. Dieter 

01232017, 09:36 PM
Post: #24




RE: Heads up for a hot new root seeking algorithm!!
(01232017 06:23 PM)Dieter Wrote:(01232017 02:22 PM)emece67 Wrote: Such Thukral algorithm is more or less useless, as it only works for zero roots. It states in the abstract that it can be used solely when the root is zero. Thus, the title f(0) = 0 is correct, the algorithm only converges to a root (yeah, with order 3 and only one function evaluation per iteration) if such root is zero. The equation Namir used was (7) in the Thukral's 2016 paper. Such equation only converges to zero roots (as stated in the abstract and conclusions). In the same paper there's a reference to another "Thukral method" (equations (19)(20)) which is the one in the 2008 paper by the same author. Equation (20) has an error (in the 2016 paper), but it is not related to equation (7), which is new. Regards. César  Information must flow. 

01232017, 10:53 PM
Post: #25




RE: Heads up for a hot new root seeking algorithm!!
I am conducting a study on the robustness of the convergence rates calculated by various mathematicians for their new rootseeking algorithms and also for wellknow rootseeking methods like Newton and Halley. I am comparing the order of convergence with actual values for various sample nonlinear functions.
Stay tuned ........... :) Namir 

01242017, 08:30 PM
(This post was last modified: 01242017 08:33 PM by emece67.)
Post: #26




RE: Heads up for a hot new root seeking algorithm!!
(01242017 07:05 PM)Mike (Stgt) Wrote: After some tests I am able to answer my question on my own. The classical Newton wins the race for square and forth root, the later even when computing it by Sqrt(Sqrt(x)). In my case, to compute 4th roots to 10000 digits the NewtonRaphson method needed around 15 iterations (30 function evaluations) whereas the OstrowskyTraub uses 8 iterations (24 function evaluations). The OT method requires also less time (75.6 ms vs 81.4 ms). I'm working in Python with the mpmath package. I've also checked the Halley method (10 iterations, 30 function evaluations, 78.7 ms, still faster than NR) and the PotraPtak method (the 3rd order one, 11 iterations, 33 functions evaluations, 101.6 ms, the slower one, curious, as this method seems quite simple). I've not found such dramatic speedups between methods as you have found, the difference between my fastest (OT) and slowest (PP) is just 1.35x. No "sliding accuracy" in my case. Regards. César  Information must flow. 

01262017, 08:21 PM
(This post was last modified: 01262017 08:30 PM by emece67.)
Post: #27




RE: Heads up for a hot new root seeking algorithm!!
These are my findings so far.
Legend:
COC is rounded to the nearest integer, except for the \(x^61\) case. Derivatives are always computed by means of the analytical expressions for them (not approximated with differences). The PotraPtak method is a 3rd order method requiring two evaluations for f(x) and one for its derivative. The KungTraub method is the 8th order one. Although I expected bigger differences between different methods, the OstrowskyTraub method seems a winner to me. Regards. César  Information must flow. 

01272017, 10:02 PM
Post: #28




RE: Heads up for a hot new root seeking algorithm!!
And these are using an approximate derivative. I think that the OstrowskyTraub method wins again.
Regards. COC_ad.pdf (Size: 21.26 KB / Downloads: 8) César  Information must flow. 

01282017, 12:09 AM
Post: #29




RE: Heads up for a hot new root seeking algorithm!!
Polishing my study before I post it on the web. Hang in there.
Namir 

02012017, 10:03 PM
(This post was last modified: 02012017 10:15 PM by Namir.)
Post: #30




RE: Heads up for a hot new root seeking algorithm!!
(02012017 05:39 PM)Mike (Stgt) Wrote: I did also invest few days to the subject but I do not publish my results as they are somehow 'unfair'. My focus is limited to solve a^2z and a^4z with enhanced precission no pocket calculator offers. The 'unfair' point is, that I compare the elapse time of 'Classical Newton' (CN) solving for sqare root with other methods, but use as x(n+1) = x(n)  f(x(n))/f'(x(n)) its evaluation a = .5 * (a + z / a). So per loop there are only three operations, a division, an addition and a multipliacation. I evaluated also CN of a^4z, Ostrowski, Grau et al., Weerakoon et al., and Homeier. All have from few more to many more operations which need more time to calculate with the enhanced settings of accurancy (10'000 and more) than the order of convergence might save. Oh wow! You use REXX and ooREXX?? I have them installed on my PC but never have the time to use them. I found 00REXX about twenty years ago! I had reviewed REXX for either BYTE magazine or Computer Language. It is a good batch language but not a number crunching language. Counting the operations adds more insight to the CPU effort. Of course you have to exclude the operations used to evaluate the target function and its derivative(s) and instead count the number of times you call them. I usually use difference approximation for the first and second derivatives. I manage to issue the same number of total function calls as the number of calls to f(x), f'(x), and f''(x) if the latter is needed. Of course some functions have derivatives that have simpler expressions and some have derivatives with more elaborate expressions. "Wish it were so simple" (Paraphrasing from the movie "Hail Caesar"). Namir 

« Next Oldest  Next Newest »

User(s) browsing this thread: 1 Guest(s)