গণনাযোগ্য ফাংশনগুলিকে কেন পুনরাবৃত্ত ফাংশন বলা হয়?


23

গণনাযোগ্যতার তত্ত্বে, গণনাযোগ্য ফাংশনগুলিকে পুনরাবৃত্ত ফাংশনও বলা হয়। কমপক্ষে প্রথম দর্শনে, আপনি প্রতিদিন-দিনের প্রোগ্রামিংয়ে (যেমন, ফাংশনগুলি যা তাদেরকে বলা হয়) এর সাথে যা কিছু বলে তার সাথে মিল নেই।

গণনার প্রেক্ষাপটে পুনরাবৃত্তির আসল অর্থ কী? এই ফাংশনগুলিকে "পুনরাবৃত্ত" বলা হয় কেন?

এটি অন্য কথায় বলতে গেলে: "পুনরাবৃত্তি" এর দুটি অর্থের মধ্যে সংযোগ কী?


2
rec-

3
তারা ঠকায়, কারণ তারা μ অপারেটরকে অন্তর্ভুক্ত করে । এটি একটি মিনিমাইজেশন অপারেটর, তবে অবশ্যই সংক্ষিপ্তকরণটির পুনরাবৃত্তির সাথে খুব কম সম্পর্ক রয়েছে। সুতরাং দেখে মনে হচ্ছে যে কেউ (ক্লেইন) ভেবেছিল যে "পুনরাবৃত্ত" ভাল লাগবে, তাই তিনি এই নামটি ব্যবহার করার জন্য একটি অজুহাত আবিষ্কার করেছিলেন। অনেক পরে, রবার্ট সোয়ার ব্যাখ্যা করেছিলেন যে "গণনাযোগ্য" আরও ভাল শোনাবে এবং "পুনরাবৃত্তি" কেবল প্রথম দিনগুলির একটি বিপণন কৌশল ছিল এবং সকলেই তাতে সম্মত হয়েছিল।
টমাস ক্লিম্পেল

3
আদিম পুনরাবৃত্তির কাজ সম্পর্কে ওয়াহট? উইকিপিডিয়া থেকে অনুলিপি করা হয়েছে তারা এবং এইচ ( এস ( ওয়াই ) , এক্স 1 , , এক্স কে ) = জি ( y ) হিসাবে সংজ্ঞায়িত হয়েছে , এইচ ( y , x 1 , h(0,x1,,xk)=f(x1,,xk)। এটি একটি ফাংশন যা নিজেকে কল করে। h(S(y),x1,,xk)=g(y,h(y,x1,,xk),x1,,xk)
হেন্ডরিক জানুয়ারী

3
@ গলোরোডন নোট করুন যে 'গণনাযোগ্যতা' এর ট্যাগ-বিবরণ (আপনি এটি এই প্রশ্নের জন্য ব্যবহার করেছেন) বলেছেন: "গণনাযোগ্যতা তত্ত্ব ওরফে পুনরাবৃত্তি তত্ত্ব"। গডেল ফাংশনকে পুনরাবৃত্ত বলে অভিহিত করেছেন , তবে এই শব্দটি বিবর্তনযোগ্য হিসাবে বিকশিত হয়েছিল । সম্ভবত আপনার মত বিভ্রান্তি এড়াতে। যে লোকেরা গণনীয়তা তত্ত্ব অধ্যয়ন করে (নিবিড়ভাবে) তার শিকড়গুলির 'সম্মান' করার জন্য পুনরাবৃত্তি তত্ত্বটি বেশি ব্যবহার করে।
আউবারন

1
কারণ তারা পুনরাবৃত্তভাবে সংজ্ঞায়িত হয়, অর্থাত্ " আরও জটিল ফাংশনগুলি পূর্বনির্ধারিত, সহজ ফাংশনগুলির ক্ষেত্রে সংজ্ঞায়িত করা হয় "
নিকোস এম।

উত্তর:


13

কিছু বেসিক ফাংশন সংজ্ঞায়িত করুন:

  • শূন্য ফাংশন

    zero:NN:x0
  • উত্তরসূরি ফাংশন

    succ:NN:xx+1
  • অভিক্ষেপ ফাংশন

pin:NnN:(x1,x2,,xn)xi

এখন থেকে আমি চিহ্নিত করতে ব্যবহার করব ( x 1 , x 2 , , এক্স এন )xn¯(x1,x2,,xn)

একটি সংজ্ঞা সংজ্ঞা:

প্রদত্ত ফাংশন

  • প্রতিটি স্বাক্ষরের সাথে এন কেএনg1,g2,,gmNkN
  • f:NmN

নিম্নলিখিত ফাংশনটি নির্মাণ করুন:

h:NkN:xk¯h(xk¯)=f(g1(xk¯),g2(xk¯),,gm(xk¯))

আদিম পুনরাবৃত্তি সংজ্ঞায়িত করুন:

প্রদত্ত ফাংশন

  • f:NkN
  • g:Nk+2N

নিম্নলিখিত (অংশবিশেষ) ফাংশনটি তৈরি করুন:

h:Nk+1N:(xk¯,y+1){f(xk¯),y+1=0g(xk¯,y,h(xk¯,y)),y+1>0

All functions that can be made using compositions and primitive recursion on basic functions, are called primitive recursive. It is called that way by definition. While a link with functions that call themselves exists, there's no need to try and link them with each other. You might consider recursion a homonym.

This definition and construction above was constructed by Gödel (a few other people were involved too) in an attempt to capture all functions that are computable i.e. there exists a Turing Machine for that function. Note that the concept of a Turing Machine was not yet described, or it was at least very vague.

(Un)fortunately, someone called Ackermann came along and defined the following function:

  • Ack:N2N
  • Ack(0,y)=y+1
  • Ack(x+1,0)=Ack(x,1)
  • Ack(x+1,y+1)=Ack(x,Ack(x+1,y))

This function is computable, but there's no way to construct it using only the constructions above! (i.e. Ack is not primitive recursive) This means that Gödel and his posse failed to capture all computable functions in their construction!

Gödel had to expand his class of functions so Ack could be constructed. He did this by defining the following:

Unbounded minimisation

  • g:NkN
  • IF [f(xk¯,y)=0 AND f(xk¯,z) is defined z<y AND f(xk¯,z)0]
    THEN
    g(xk¯)=y
    ELSE
    g(xk¯) is not defined.

This last one may be hard to grasp, but it basically means that g((x1,x2,,xk)) is the smallest root of f (if a root exists).


All functions that can be constructed with all the constructions defined above are called recursive. Again, the name recursive is just by definition, and it doesn't necessarily have correlation with functions that call themselves. Truly, consider it a homonym.

Recursive functions can be either partial recursive functions or total recursive functions. All partial recursive functions are total recursive functions. All primitive recursive functions are total. As an example of a partial recursive function that is not total, consider the minimisation of the successor function. The successor function doesn't have roots, so its minimisation is not defined. An example of a total recursive function (which uses minimisation) is Ack.

Now Gödel was able to construct the Ack function as well with his expanded class of functions. As a matter of fact, every function that can be computed by a Turing machine, can be represented by using the constructions above and vice versa, every construction can be represented by a Turing machine.

If you're intrigued, you could try to make Gödel's class bigger. You can try to define the 'opposite' of unbounded minimisation. That is, unbounded maximisation i.e. the function that finds the biggest root. However, you may find that computing that function is hard (impossible). You can read into the Busy Beaver Problem, which tries to apply unbounded maximisation.


4
I know realise the given definitions don't really answer the question, but my answer describes the evolution of recursion/computability theory, kind of. Might be worth a read.
Auberon

I like it, thanks for your efforts :-)
Golo Roden

In "if h((x1,x2,...,xk),0)=f((x1,x2,...,xk))", I think you mean h((x1,x2,...,xk,0)). Also, there is no then clause prior to the next bullet point's else clause.
Eric Towers

2
Afaik, this is subtly wrong. The set if μ-recursive functions is called the set of partially recursive functions whereas recursive functions are always total. That's why the set of all total functions (resp. languages that can be decided) is called R.
Raphael

1
There are quite a few incorrect statements in your answer. You should not make up history for an answer.
Kaveh

17

The founders of computability theory were mathematicians. They founded what is now called computability theory before there was any computers. What was the way mathematicians defined functions that could be computed? By recursive definitions!

So there were recursive function before there were any other model of computation like Turing machines or lambda calculus or register machines. So people referred to these function as recursive functions. The fact that they turned out to be exactly what Turing machines and other models can compute is a later event (mostly proven by Kleene).

We have the simple definition of a recursive function which is now called primitive recursive function. There were not general enough (e.g. Ackermann's function) so people developed more general notions like μ-recursive functions and Herbrand-Gödel general recursive functions that did capture all computable functions (assuming the Church's thesis). Church claimed that his model of lambda calculus captured all computable functions. Many people, and in particular Gödel, were not convinced that these capture all functions that can be computed. Until Turing's analysis of computation and introduction of his machine model.

The name of the field used to recursion theory. However there has been a successful push in recent decades to change the name to something more appealing from recursion theory to something more computer sciency (vs. mathy). As a result the field is now called computability theory. However if you look at books, papers, conferences, etc. in the early decades they are called recursion theory and not computability theory. Even the title of Soare's own 1987 book (who was the main person behind the push to change the name to computability theory) is "Recursively Enumerable Sets and Degrees".

If you want to know more about the history a fun and good place to read about it is the first chapter of Classical Recursion Theory by Odifreddi.


7

Robert Soare wrote an essay about this issue. According to him, the term (general) recursive functions was coined by Gödel, who defined them using some sort of mutual recursion. The name stuck, though later on other equivalent definitions were found.

For more information, I recommend Soare's essay.


0

instead of putting a long comment decided to add an answer:

Because they are defined recursively, i.e "more complex functions are defined in terms of previously defined, simpler functions"

This kind of iterative or incremental procedure creates well-defined functions (in the mathematical sense)

This is the meaning of recursiveness in mathematical parlance. See below how this relates to recursion in programming parlance.

Compare this procedure with techniques and methods like (mathematical) induction which is also an example of recursiveness in mathematics.

Programming has a mathematical vein as well as an engineering one.

This (usualy constructive) procedure is also refered as "bootstrapping" in Operating Systems parlance.

However a runtime recursion of the same function (i.e caling itself during its runtime), since it must (hmm, should) happen on already computed values (or arguments), or in other words, in the part of the result set already computed, is also recursive in the above sense, i.e "defined w.r.t previously defined functions (and their values)"

Else is not well-defined, and leads to such things like Stack Overflow :))))

To give a further example from Operating Systems, a runtime recursion (calling itself) can be taken as the analog of an operating system rebooting after a certain update (e.g core update). Many OSes do the following procedure:

  1. an initial boot to load low-level routines (e.g I/O)
  2. do the necessary updates (using the low-level routines)
  3. re-boot (effectively, re-calling itself), but this time loading the more complex routines (or even the whole system)

Auberon's beautiful answer demonstrates a procedure of this kind in more detail.

আমাদের সাইট ব্যবহার করে, আপনি স্বীকার করেছেন যে আপনি আমাদের কুকি নীতি এবং গোপনীয়তা নীতিটি পড়েছেন এবং বুঝতে পেরেছেন ।
Licensed under cc by-sa 3.0 with attribution required.