Programmimg Pearls (1985)

(dl.acm.org)

106 points | by edent 9 days ago

10 comments

  • tkhattra 8 days ago
    bentley revised and updated some of his "programming pearls" columns and published them in 2 editions of books published circa 1986 and 2000. they are among my favorite computer science books, and a great companion to more theoretical data structures, algorithms, and software engineering texts. a great mix of theory, engineering, rules of thumb, insight, creativity etc., all presented lucidly. highly recommended!
    • vram22 8 days ago
      Yes, those two books are Programming Pearls and More Programming Pearls. I have both, and they are good. In one of them there is a section or chapter with a few examples of advanced uses of awk, involving awk's associative arrays, to do some non-trivial work in very few lines of code[1]. He shows or says that the equivalent programs in C would be much longer and harder to write.

      [1] Even more so than in most other common examples where awk programs are shorter than those in other languages.

  • gregfjohnson 7 days ago
    If this is the first time you’ve heard of Jon Bentley’s “Bumper sticker computer science”, you are in for a rare treat! It is a combination of the hilarious and the deeply insightful. An absolute joy, and a must-read for computer scientists and programmers. It is a distillation of aphorisms and wisdom from an entire generation of many of the world’s best minds in our discipline.
  • satyanash 8 days ago
    > Allocate four digits for the year part of the date: a new millenium is coming -- David Martin, Norristown Pennsylvania

    2038 is 15 years away right now; this article was published 15 years before 2000.

    • bombcar 8 days ago
      Even in 1985 you could foresee wanting to reference something after 2000 - terms for a thirty year loan, for example.

      Harder to see reasons past 2038 back in 1970.

      • nerdponx 7 days ago
        > Harder to see reasons past 2038 back in 1970

        I still find myself confused at this one. Did people in 1970s not imagine that any particular program or database would "survive" for 30+ years? Was the expectation that programs written in the future could set the epoch to a later date?

        Now that I think about it, I also find it interesting that the epoch date isn't configurable in any system that I'm aware of.

        • Gordonjcp 7 days ago
          Unix timestamps can represent dates *before* 1970, by using negative numbers.

          Dennis Ritchie has said that he picked signed 32-bit values with an epoch of 1/1/1970 because that would comfortably represent dates exceeding his life - before he was born until after his (expected) death.

        • p_l 7 days ago
          Nobody expected Unix timestamp to survive so long when it was designed, and norm was to use different format in databases that would care about dates beyond 2038
          • bombcar 7 days ago
            And memory was very precious. Why waste bits that would always be zero?
            • p_l 7 days ago
              Not just bits - unix timestamp ossified as 32bit value on PDP-11, and handling longer timestamp would involve a lot more code to implement arithmetic on larger than 32bit values, plus also would complicate C language
        • guenthert 7 days ago
          All computer evolution up to that day happened in the 30 years prior, so how on earth could they possibly expect that their system would last such a long time?

          I do recall from the time frame mid-eighthies to mid-nineties that three year old computers were obsolete. Ten year old stuff was ancient history. Never would I have imagined to be able to use Linux for 30 years.

          • Throwawayaerlei 7 days ago
            IBM's promises, largely delivered to this day, that their System/360 computer hardware and future generations would be compatible, as was that line of computers except for the lowest end special case? All the effort IBM put into emulation of their older machines in the microcoded models (all but two high end ones)?

            Some time before then it was realized we had a "software crisis" as outsiders put it, and IBM realized they were spending too much money supporting many different designs, some evolutionary with minor changes?

            Many ideas behind systems like Linux go back to the 1960s as well, see Multics, the primary inspiration for UNIX but ultimately a closed source dead end.

            For the original question we have to realize how very small many old computers were. UNIX started on small DEC minicomputers, it would have made almost no sense at all back then to allocate more than 32 bits when your maximum data address space was 56KiB and your CPU was 16 bit (the PDP-11 family where UNIX became big following an incompatible DEC predecessor).

        • varjag 7 days ago
          It's a worthy exercise to code for a while on a vintage machine with a hundred or so kilowords memory. Puts things into perspective.
    • throwawaytemp29 7 days ago
      Y10K bug! Only 4 digit dates.
  • biryani_chicken 8 days ago
    It's also in archive.org [0] with better quality and several formats for download.

    [0] https://archive.org/details/ProgrammingPearls2ndEditionJonBe...

    • tkhattra 7 days ago
      i don't think that's the text of the entire book - it looks like a scrape of the book's former website (www.programmingpearls.com).
  • embit 7 days ago
    John B. on quick sort. Just awesome explanation

    https://m.youtube.com/watch?v=QvgYAQzg1z8

    • bhrgunatha 7 days ago
      The analysis of comparisons was both brilliant and illuminating.
  • watersb 8 days ago
    A series of exercises from 1970s, so exercises tend to be good Perl or Bash one-liners these days.

    But! The real point of this book is How to Ask Questions.

    Anyone who has written an entire AI-based inference engine, when the customer wanted a spreadsheet, will get a kick out of this.

    And those of us who haven't (yet) will gain some insight that might prove useful later on.

  • macintux 8 days ago
    I thought the title here had a typo, but no, indeed, the typo is at the ACM website (but not the paper itself).

    Some choice tidbits in there.

    > The job’s not over until the paperwork’s done.

  • NorSoulx 7 days ago
    I co-designed and developed one of the early SMS content platforms in Norway back in 2000. I had bought and read Programming Pearls, and used ideas from this book to implement number portability routing/handling when it was introduced for mobile operator phone numbers in Norway.
  • eimrine 7 days ago
    Why to avoid asymmetry?
    • ezekiel68 7 days ago
      This is only an opinion --but it's the opinion of someone who has been involved in software engineering since the 1990s:

      Since this quote is in the section of general advice, I feel this advice applies mainly to the analysis and design of systems. Here's an example: You want to make a command line program (or it could be an API) that can behave differently, depending on certain conditions. So you have an analysis and design problem, which is: how do you cover these different scenarios (flexibility) without making the program/api too difficult to use (complexity)?

      If you hard-code the choices into the system, it gets a part of the job done, but there is no flexibility and so your system will be useless to those who need the outcome to match certain common use cases. But, on the other hand, if you require users to type in a long list of parameters every time they use the system (complexity), users will avoid this solution because it is simply too difficult to use for the most common scenarios.

      Symmetry in this sense means striking the right balance among the aspects of usefulness and simplicity so that your system will match the needs of its intended users. In my hypothetical case, the programmer might provide a config file which lists comments showing the default values for the config options. If a user needs to tweak this behavior slightly from the defaults, they can set the config values as it suits them in the config file once, rather than needing to remember to type the overrides in as command-line params each time.

      Again -- this is one simple example. When defining a data model (a different realm where symmetry applies), it is common to take many iterations to "get it right" so that the system can properly reflect a real-world problem, such as how to represent that 'many students take many classes at the college'. Early efforts tend to be "asymmetric" in the sense that they fail to capture key relationships between important concepts by making the data model too "flat" or simplistic. Further modeling efforts, though, may make the opposite mistake by trying too hard to faithfully capture every possible corner case, including those which either seldom occur or have no practical value (such as recording whether any instructors took leaves of abscence during a semester, when the point of the system is to capture student attendance and grades).

    • toolslive 7 days ago
      "asymmetry" typically involves exceptional state/behaviour. It's an extra cost while you're building it, and it's an extra cost while maintaining it. For example, a customer wants something special but it will only be needed by that customer.
    • edent 7 days ago
      I was wondering about that. I think it is about having the ability to go A->B without being able to go B->A.

      But, as with many of the Pearls, it is heavily context dependent. And if you aren't well versed in "primitive" computing paradigms, you'll be confused.

    • epgui 7 days ago
      It's a bit hard to explain this one, but symmetry is a form of simplicity. This might make more sense to a physicist or mathematician.
  • fernly 8 days ago
    undefined