Skip to main content

Harvard Professor Explains Algorithms in 5 Levels of Difficulty

From the physical world to the virtual world, algorithms are seemingly everywhere. David J. Malan, Professor of Computer Science at Harvard University, has been challenged to explain the science of algorithms to 5 different people; a child, a teen, a college student, a grad student, and an expert. Director: Wendi Jonassen Director of Photography: Zach Eisen Editor: Louville Moore Host: David J. Malan Guests: Level 1: Addison Vincent Level 2: Lexi Kemmer Level 3: Patricia Guirao Level 4: Mahi Shafiullah Level 5: Chris Wiggins Creative Producer: Maya Dangerfield Line Producer: Joseph Buscemi Associate Producer: Paul Gulyas; Kameryn Hamilton Production Manager: D. Eric Martinez Production Coordinator: Fernando Davila Casting Producer: Vanessas Brown; Nicholas Sawyer Camera Operator: Brittany Berger Gaffer: Gautam Kadian Sound Mixer: Lily Van Leeuwen Production Assistant: Ryan Coppola Hair & Make-Up: Yev Wright-Mason Post Production Supervisor: Alexa Deutsch Post Production Coordinator: Ian Bryant Supervising Editor: Doug Larsen Assistant Editor: Lauren Worona

Released on 11/08/2023

Transcript

Hello world.

My name is David J. Malan

and I'm a professor of computer science

at Harvard University.

Today, I've been asked to explain algorithms

in five levels of increasing difficulty.

Algorithms are important

because they really are everywhere,

not only in the physical world,

but certainly in the virtual world as well.

And in fact, what excites me about algorithms

is that they really represent an opportunity

to solve problems.

And I dare say, no matter what you do in life,

all of us have problems to solve.

So, I'm a computer science professor,

so I spend a lot of time with computers.

How would you define a computer for them?

Well, a computer is electronic,

like a phone but it's a rectangle,

and you can type like tick, tick, tick.

And you work on it.

Nice. Do you know any of the parts

that are inside of a computer?

No.

Can I explain a couple of them to you?

Yeah.

So, inside of every computer is some kind of brain

and the technical term for that is CPU,

or central processing unit.

And those are the pieces of hardware

that know how to respond to those instructions.

Like moving up or down, or left or right,

knows how to do math like addition and subtraction.

And then there's at least one other type of

hardware inside of a computer called memory

or RAM, if you've heard of this?

I know of memory because you have to memorize stuff.

Yeah, exactly.

And computers have even different types of memory.

They have what's called RAM, random access memory,

which is where your games, where your programs

are stored while they're being used.

But then it also has a hard drive,

or a solid state drive, which is where your data,

your high scores, your documents,

once you start writing essays and stories in the future.

It stays there.

Stays permanently.

So, even if the power goes out,

the computer can still remember that information.

It's still there because

the computer can't just like delete all of the words itself.

Hopefully not.

Because your fingers could only do that.

Like you have to use your finger to delete

all of the stuff. Exactly.

You have to write.

Yeah, have you heard of an algorithm before?

Yes. Algorithm is a list of instructions to tell people

what to do or like a robot what to do.

Yeah, exactly.

It's just step by step instructions for doing something,

for solving a problem, for instance.

Yeah, so like if you have a bedtime routine,

then at first you say, I get dressed, I brush my teeth,

I read a little story, and then I go to bed.

All right.

Well how about another algorithm?

Like what do you tend to eat for lunch?

Any types of sandwiches you like?

I eat peanut butter.

Let me get some supplies from the cupboard here.

So, should we make an algorithm together?

Yeah.

Why don't we do it this way?

Why don't we pretend like I'm a computer

or maybe I'm a robot, so I only understand your instructions

and so I want you to feed me, no pun intended, an algorithm.

So, step-by-step instructions for solving this problem.

But remember, algorithms, you have to be precise,

you have to give...

The right instructions.

[David] The right instructions.

Just do it for me. So, step one was what?

Open the bag.

[David] Okay. Opening the bag of bread.

[David] Stop. Now what?

Grab the bread and put it on the plate.

[David] Grab the bread and put it on the plate.

Take all the bread back and put it back in there.

[David laughing]

So, that's like an undo command.

Yeah.

Little control Z? Okay.

Take one bread and put it on the plate.

Okay.

Take the lid off the peanut butter.

[David] Okay, take the lid off the peanut butter.

Put the lid down.

[David] Okay. Take the knife.

[David] Take the knife.

[Addison] Put the blade inside the peanut butter

and spread the peanut butter on the bread.

I'm going to take out some peanut butter

and I'm going to spread the peanut butter on the bread.

I put a lot of peanut butter on

because I love peanut butter.

Oh, apparently. I thought I was messing with you here...

No, no it's fine.

But I think you're apparently happy with this.

[Addison] Put the knife down,

and then grab one bread and put it on top

of the second bread, sideways.

Sideways.

Like put it flat on.

Oh, flat ways, okay.

[Addison] And now, done. You're done with your sandwich.

Should we take a delicious bite?

Yep. Let's take a bite.

[David] Okay, here we go.

What would the next step be here?

Clean all this mess up.

[David laughing]

Clean all this mess up, right.

We made an algorithm, step by step instructions

for solving some problem.

And if you think about now,

how we made peanut butter and jelly sandwiches,

sometimes we were imprecise and you didn't give me

quite enough information to do the algorithm correctly,

and that's why I took out so much bread.

Precision, being very, very correct with your instructions

is so important in the real world

because for instance, when you're using the worldwide web

and you're searching for something on Google or Bing...

You want to do the right thing.

[David] Exactly.

So, like if you type in just Google,

then you won't find the answer to your question.

Pretty much everything we do in life is an algorithm,

even if we don't use that fancy word to describe it.

Because you and I are sort of following instructions

either that we came up with ourselves

or maybe our parents told us how to do these things.

And so, those are just algorithms.

But when you start using algorithms in computers,

that's when you start writing code.

[upbeat music]

What do you know about algorithms?

Nothing really, at all honestly.

I think it's just probably a way to store information

in computers.

And I dare say, even though you might not have

put this word on it, odds are you executed as a human,

multiple algorithms today even before you came here today.

Like what were a few things that you did?

I got ready.

Okay. And get ready. What does that mean?

Brushing my teeth, brushing my hair.

[David] Okay.

Getting dressed.

Okay, so all of those, frankly, if we really

dove more deeply, could be broken down into

step-by-step instructions.

And presumably your mom, your dad, someone in the past

sort of programmed you as a human to know what to do.

And then after that, as a smart human,

you can sort of take it from there

and you don't need their help anymore.

But that's kind of what we're doing

when we program computers.

Something maybe even more familiar nowadays,

like odds are you have a cell phone.

Your contacts or your address book.

But let me ask you why that is.

Like why does Apple or Google or anyone else

bother alphabetizing your contacts?

I just assumed it would be easier to navigate.

What if your friend happened to be at the very bottom

of this randomly organized list?

Why is that a problem? Like he or she's still there.

I guess it would take a while to get to

while you're scrolling.

That, in of itself, is kind of a problem

or it's an inefficient solution to the problem.

So, it turns out that back in my day,

before there were cell phones, everyone's numbers

from my schools were literally printed in a book,

and everyone in my town and my city, my state

was printed in an actual phone book.

Even if you've never seen this technology before,

how would you propose verbally to find John

in this phone book? Or I would just flip through

and just look for the J's I guess.

Yeah. So, let me propose that we start that way.

I could just start at the beginning

and step by step I could just look at each page,

looking for John, looking for John.

Now even if you've never seen this here technology before,

it turns out this is exactly what your phone could be doing

in software, like someone from Google or Apple or the like,

they could write software that uses a technique

in programming known as a loop,

and a loop, as the word implies,

is just sort of do something again and again.

What if instead of starting from the beginning

and going one page at a time,

what if I, or what if your phone goes like two pages

or two names at a time?

Would this be correct do you think?

Well you could skip over John, I think.

In what sense?

If he's in one of the middle pages that you skipped over.

Yeah, so sort of accidentally and frankly

with like 50/50 probability,

John could get sandwiched in between two pages.

But does that mean I have to throw

that algorithm out altogether?

Maybe you could use that strategy until you get close

to the section and then switch to going one by one.

Okay, that's nice.

So, you could kind of like go twice as fast

but then kind of pump the brakes as you near your exit

on the highway, or in this case near the J section

of the book.

Exactly.

And maybe alternatively, if I get to like

A, B, C, D, E, F, G, H, I, J, K,

if I get to the K section,

then I could just double back like one page

just to make sure John didn't get sandwiched

between those pages.

So, the nice thing about that second algorithm

is that I'm flying through the phone book

like two pages at a time.

So, 2, 4, 6, 8, 10, 12.

It's not perfect, it's not necessarily correct

but it is if I just take one extra step.

So, I think it's fixable,

but what your phone is probably doing

and frankly what I and like my parents and grandparents

used to do back in the day was we'd probably go roughly

to the middle of the phone book here,

and just intuitively, if this is an alphabetized phone book

in English, what section am I probably going to

find myself in roughly?

K?

Okay. So, I'm in the K section.

Is John going to be to the left or to the right?

To the left.

Yeah.

So, John is going to be to the left or the right

and what we can do here, though your phone

does something smarter, is tear the problem in half,

throw half of the problem away,

being left with just 500 pages now.

But what might I next do?

I could sort of naively just start at the beginning again,

but we've learned to do better.

I can go roughly to the middle here.

And you can do it again. Yeah, exactly.

So, now maybe I'm in the E section,

which is a little to the left.

So, John is clearly going to be to the right,

so I can again tear the problem poorly in half,

throw this half of the problem away,

and I claim now that if we started with a thousand pages,

now we've gone to 500, 250,

now we're really moving quickly.

Yeah.

[David] And so, eventually I'm hopefully dramatically

left with just one single page

at which point John is either on that page

or not on that page, and I can call him.

Roughly how many steps might this third algorithm take

if I started with a thousand pages

then went to 500, 250, 125,

how many times can you divide 1,000 in half? Maybe?

10.

That's roughly roughly 10.

Because in the first algorithm,

looking again for someone like Zoe in the worst case

might have to go all the way through a thousand pages.

But the second algorithm you said was 500,

maybe 501, essentially the same thing.

So, twice as fast.

But this third and final algorithm is sort of fundamentally

faster because you're sort of dividing and conquering it

in half, in half, in half,

not just taking one or two bites out of it out of a time.

So, this of course is not how we used to use phone books

back in the day since otherwise they'd be single use only.

But it is how your phone is actually searching for Zoe,

for John, for anyone else, but it's doing it in software.

Oh, that's cool.

So, here we've happened to focus on searching algorithms,

looking for John in the phone book.

But the technique we just used

can indeed be called divide and conquer,

where you take a big problem and you divide and conquer it,

that is you try to chop it up into smaller,

smaller, smaller pieces.

A more sophisticated type of algorithm,

at least depending on how you implement it,

something known as a recursive algorithm.

Recursive algorithm is essentially an algorithm

that uses itself to solve the exact same problem

again and again, but chops it smaller, and smaller,

and smaller ultimately.

[upbeat music]

Hi, my name's Patricia.

Patricia, nice to meet you.

Where are you a student at?

I am starting my senior year now at NYU.

Oh nice. And what have you been studying

the past few years?

I studied computer science and data science.

If you were chatting with a non-CS,

non-data science friend of yours,

how would you explain to them what an algorithm is?

Some kind of systematic way of solving a problem,

or like a set of steps to kind of solve

a certain problem you have.

So, you probably recall learning topics

like binary search versus linear search, and the like.

Yeah.

So, I've come here complete with a

actual chalkboard with some magnetic numbers on it here.

How would you tell a friend to sort these?

I think one of the first things we learned was

something called bubble sort.

It was kind of like focusing on smaller bubbles

I guess I would say of the problem,

like looking at smaller segments rather than

the whole thing at once.

What is I think very true about what you're hinting at

is that bubble sort really focuses on local, small problems

rather than taking a step back trying to fix

the whole thing, let's just fix the obvious problems

in front of us. So, for instance, when we're trying to get

from smallest to largest,

and the first two things we see are eight followed by one,

this looks like a problem 'cause it's out of order.

So, what would be the simplest fix,

the least amount of work we can do

to at least fix one problem?

Just switch those two numbers

'cause one is obviously smaller than eight.

Perfect. So, we just swap those two then.

You would switch those again.

Yeah, so that further improves the situation

and you can kind of see it,

that the one and the two are now in place.

How about eight and six?

[Patricia] Switch it again.

Switch those again. Eight and three?

Switch it again.

[fast forwarding]

And conversely now the one and the two are closer to,

and coincidentally are exactly where we want them to be.

So, are we done?

No.

Okay, so obviously not, but what could we do now

to further improve the situation?

Go through it again but you don't need

to check the last one anymore because we know

that number is bubbled up to the top.

Yeah, because eight has indeed bubbled all the way

to the top. So, one and two?

[Patricia] Yeah, keep it as is.

Okay, two and six?

[Patricia] Keep it as is.

Okay, six and three?

Then you switch it.

Okay, we'll switch or swap those.

Six and four?

[Patricia] Swap it again.

Okay, so four, six and seven?

[Patricia] Keep it.

Okay. Seven and five?

[Patricia] Swap it.

Okay. And then I think per your point,

we're pretty darn close.

Let's go through once more.

[Patricia] One and two? Keep it.

[Patricia] Two three? Keep it.

[Patricia] Three four? Keep it.

[Patricia] Four six? Keep it.

Six five?

[Patricia] And then switch it.

All right, we'll switch this. And now to your point,

we don't need to bother with the ones

that already bubbled their way up.

Now we are a hundred percent sure it's sorted.

Yeah.

And certainly the search engines of the world,

Google and Bing and so forth,

they probably don't keep webpages in sorted order

'cause that would be a crazy long list

when you're just trying to search the data.

But there's probably some algorithm underlying what they do

and they probably similarly, just like we,

do a bit of work upfront to get things organized

even if it's not strictly sorted in the same way

so that people like you and me and others

can find that same information.

So, how about social media?

Can you envision where the algorithms are in that world?

Maybe for example like TikTok, like the For You page,

'cause those are like recommendations, right?

It's sort of like Netflix recommendations

except more constant because it's just every video

you scroll, it's like that's a new recommendation basically.

And it's based on what you've liked previously,

what you've saved previously, what you search up.

So, I would assume there's some kind of algorithm there

kind of figuring out like what to put on your For You page.

Absolutely. Just trying to keep you presumably

more engaged.

So, the better the algorithm is,

the better your engagement is,

maybe the more money the company then makes on the platform

and so forth.

So, it all sort of feeds together.

But what you're describing really is more

artificially intelligent, if I may,

because presumably there's not someone at TikTok

or any of these social media companies saying,

If Patricia likes this post, then show her this post.

If she likes this post, then show her this other post.

Because the code would sort of grow infinitely long

and there's just way too much content for a programmer

to be having those kinds of conditionals,

those decisions being made behind the scenes.

So, it's probably a little more artificially intelligent.

And in that sense you have topics like neural networks,

and machine learning which really describe

taking as input things like what you watch,

what you click on, what your friends watch,

what they click on, and sort of trying to infer

from that instead, what should we show Patricia

or her friends next?

Oh, okay. Yeah. Yeah.

That makes like the distinction more...

Makes more sense now.

Nice. Yeah.

[upbeat music]

I am currently a fourth year PhD student at NYU.

I do robot learning, so that's half and half

robotics and machine learning.

Sounds like you've dabbled with quite a few algorithms.

So, how does one actually research algorithms

or invent algorithms?

The most important way is just trying to think about

inefficiencies, and also think about connecting threads.

The way I think about it is that algorithm for me

is not just about the way of doing something,

but it's about doing something efficiently.

Learning algorithms are practically everywhere now.

Google, I would say for example,

is learning every day about like,

Oh what articles, what links might be better than others?

And re-ranking them.

There are recommender systems all around us, right?

Like content feeds and social media,

or you know, like YouTube or Netflix.

What we see is in a large part determined by this kind of

learning algorithms.

Nowadays there's a lot of concerns

around some applications of machine learning

like deep fakes where it can kind of learn how I talk

and learn how you talk and even how we look,

and generate videos of us.

We're doing this for real, but you could imagine

a computer synthesizing this conversation eventually.

Right.

But how does it even know what I sound like

and what I look like, and how to replicate that?

All of this learning algorithms that we talk about, right?

A lot, like what goes in there is just

lots and lots of data.

So, data goes in, something else comes out.

What comes out is whatever objective function

that you optimize for.

Where is the line between algorithms that

play games with and without AI?

I think when I started off my undergrad,

the current AI machine learning

was not very much synonymous.

Okay.

And even in my undergraduate, in the AI class,

they learned a lot of classical algorithms for game plays.

Like for example, the A star search, right?

That's a very simple example of how you can play a game

without having anything learned.

This is very much, oh you are at a game state,

you just search down, see what are the possibilities

and then you pick the best possibility that it can see,

versus what you think about when you think about,

ah yes, gameplay like the alpha zero for example,

or alpha star, or there are a lot of, you know,

like fancy new machine learning agents that are

even learning very difficult games like Go.

And those are learned agents, as in they are getting better

as they play more and more games.

And as they get more games, they kind of

refine their strategy based on the data that I've seen.

And once again, this high level abstraction

is still the same.

You see a lot of data and you'll learn from that.

But the question is what is objective function

that you're optimizing for?

Is it winning this game?

Is it forcing a tie or is it, you know,

opening a door in a kitchen?

So, if the world is very much focused on supervised,

unsupervised reinforcement learning now,

what comes next five, ten years, where is the world going?

I think that this is just going to be more and more,

I don't want to use the word encroachment,

but that's what it feels like of algorithms

into our everyday life.

Like even when I was taking the train here, right?

The trains are being routed with algorithms,

but this has existed for you know, like 50 years probably.

But as I was coming here, as I was checking my phone,

those are different algorithms,

and you know, they're kind of getting all around us,

getting there with us all the time.

They're making our life better most places, most cases.

And I think that's just going to be a continuation

of all of those.

And it feels like they're even in places

you wouldn't expect, and there's just so much data

about you and me and everyone else online

and this data is being mined and analyzed,

and influencing things we see and hear it would seem.

So, there is sort of a counterpoint which might be good

for the marketers, but not necessarily good for you and me

as individuals.

We are human beings, but for someone

we might be just a pair of eyes who are

carrying a wallet, and are there to buy things.

But there is so much more potential for these algorithms

to just make our life better without

changing much about our life.

[upbeat music]

I'm Chris Wiggins. I'm an associate professor

of Applied Mathematics at Columbia.

I'm also the chief data scientist of the New York Times.

The data science team at the New York Times

develops and deploys machine learning

for newsroom and business problems.

But I would say the things that we do mostly, you don't see,

but it might be things like personalization algorithms,

or recommending different content.

And do data scientists, which is rather distinct

from the phrase computer scientists.

Do data scientists still think in terms of algorithms

as driving a lot of it?

Oh absolutely, yeah.

In fact, so in data science and academia,

often the role of the algorithm is

the optimization algorithm that helps you find the best

model or the best description of a data set.

And data science and industry, the goal,

often it's centered around an algorithm

which becomes a data product.

So, a data scientist in industry might be

developing and deploying the algorithm,

which means not only understanding the algorithm

and its statistical performance,

but also all of the software engineering

around systems integration, making sure that that algorithm

receives input that's reliable and has output that's useful,

as well as I would say the organizational integration,

which is how does a community of people

like the set of people working at the New York Times

integrate that algorithm into their process?

Interesting. And I feel like AI based startups

are all the rage and certainly within academia.

Are there connections between AI

and the world of data science?

Oh, absolutely.

The algorithms that they're in,

can you connect those dots for...

You're right that AI as a field has really exploded.

I would say particularly many people experienced a ChatBot

that was really, really good.

Today, when people say AI,

they're often thinking about large language models,

or they're thinking about generative AI,

or they might be thinking about a ChatBot.

One thing to keep in mind is a ChatBot is a special case

of generative AI, which is a special case of using

large language models, which is a special case of using

machine learning generally,

which is what most people mean by AI.

You may have moments that are what John McCarthy called,

Look Ma, no hands, results,

where you do some fantastic trick and you're not quite sure

how it worked.

I think it's still very much early days.

Large language models is still in the point of

what might be called alchemy and that people are building

large language models without a real clear,

a priori sense of what the right design is

for a right problem.

Many people are trying different things out,

often in large companies where they can afford

to have many people trying things out,

seeing what works, publishing that,

instantiating it as a product.

And that itself is part of the scientific process

I would think too.

Yeah, very much. Well, science and engineering,

because often you're building a thing

and the thing does something amazing.

To a large extent we are still looking for

basic theoretical results around why

deep neural networks generally work.

Why are they able to learn so well?

They're huge, billions of parameter models

and it's difficult for us to interpret

how they're able to do what they do.

And is this a good thing, do you think?

Or an inevitable thing that we, the programmers,

we, the computer scientists, the data scientists

who are inventing these things,

can't actually explain how they work?

Because I feel like friends of mine in industry,

even when it's something simple and relatively familiar

like auto complete, they can't actually tell me

why that name is appearing at the top of the list.

Whereas years ago when these algorithms were more

deterministic and more procedural,

you could even point to the line that made that name

[Chris] bubble up to the top. Absolutely.

So, is this a good thing, a bad thing,

that we're sort of losing control perhaps in some sense

of the algorithm?

It has risks.

I don't know that I would say that it's good or bad,

but I would say there's lots of scientific precedent.

There are times when an algorithm works really well

and we have finite understanding of why it works

or a model works really well

and sometimes we have very little understanding

of why it works the way it does.

In classes I teach, certainly spend a lot of time on

fundamentals, algorithms that have been taught in classes

for decades now, whether it's binary search,

linear search, bubble sorts, selection sort or the like,

but if we're already at the point where I can pull up

chat GPT, copy paste a whole bunch of numbers or words

and say, Sort these for me,

does it really matter how Chat GPT is sorting it?

Does it really matter to me as the user

how the software is sorting it?

Do these fundamentals become more dated and less important

do you think?

Now you're talking about the ways in which code

and computation is a special case of technology, right?

So, for driving a car, you may not necessarily need

to know much about organic chemistry,

even though the organic chemistry is how the car works.

So, you can drive the car and use it in different ways

without understanding much about the fundamentals.

So, similarly with computation, we're at a point

where the computation is so high level, right?

You can import psychic learn and you can go from zero

to machine learning in 30 seconds.

It's depending on what level you want to understand

the technology, where in the stack, so to speak,

it's possible to understand it and make wonderful things

and advance the world without understanding it

at the particular level of somebody who actually might have

originally designed the actual optimization algorithm.

I should say though, for many of the optimization

algorithms, there are cases where an algorithm

works really well and we publish a paper,

and there's a proof in the paper,

and then years later people realize

actually that proof was wrong and we're really

still not sure why that optimization works,

but it works really well or it inspires people

to make new optimization algorithms.

So, I do think that the goal of understanding algorithms

is loosely coupled to our progress

and advancing grade algorithms, but they don't always

necessarily have to require each other.

And for those students especially,

or even adults who are thinking of now steering into

computer science, into programming,

who were really jazzed about heading in that direction

up until, for instance, November of 2022,

when all of a sudden for many people

it looked like the world was now changing

and now maybe this isn't such a promising path,

this isn't such a lucrative path anymore.

Are LLMs, are tools like Chat GPT reason not to perhaps

steer into the field?

Large language models are a particular architecture

for predicting, let's say the next word,

or a set of tokens more generally.

The algorithm comes in when you think about

how is that LLM to be trained or also how to be fine tuned.

So, the P of GPT is a pre-trained algorithm.

The idea is that you train a large language model

on some corpus of text, could be encyclopedias,

or textbooks, or what have you.

And then you might want to fine tune that model

around some particular task or

some particular subset of texts.

So, both of those are examples of training algorithms.

So, I would say people's perception

of artificial intelligence has really changed a lot

in the last six months, particularly around November of 2022

when people experienced a really good ChatBot.

The technology though had been around already before.

Academics had already been working with Chat GPT three

before that and GPT two and GPT one.

And for many people it sort of opened up this conversation

about what is artificial intelligence

and what could we do with this?

And what are the possible good and bad, right?

Like any other piece of technology.

Kranzburg's first law of technology,

technology is neither good, nor bad, nor is it neutral.

Every time we have some new technology,

we should think about it's capabilities

and the good, and the possible bad.

[David] As with any area of study,

algorithms offer a spectrum from the most basic

to the most advanced.

And even if right now, the most advanced of those algorithms

feels out of reach because you just

don't have that background,

with each lesson you learn, with each algorithm you study,

that end game becomes closer and closer

such that it will, before long, be accessible to you

and you will be at the end of that most advanced spectrum.

Up Next