"Give a man a fish and you feed him for a day. Teach a man to fish and you feed him for a lifetime." | ||
--Chinese proverb |
According to Wikipedia, computer programming is:
"...a process that leads from an original formulation of a computing problem to executable computer programs. Programming involves activities such as analysis, developing understanding, generating algorithms, verification of requirements of algorithms including their correctness and resources consumption, and implementation (commonly referred to as coding) of algorithms in a target programming language".
In a nutshell, coding is telling a computer to do something using a language it understands.
Computers are very powerful tools, but unfortunately, they can't think for themselves. So they need to be told everything. They need to be told how to perform a task, how to evaluate a condition to decide which path to follow, how to handle data that comes from a device such as the network or a disk, and how to react when something unforeseen happens, say, something is broken or missing.
You can code in many different styles and languages. Is it hard? I would say "yes" and "no". It's a bit like writing. Everybody can learn how to write, and you can too. But what if you wanted to become a poet? Then writing alone is not enough. You have to acquire a whole other set of skills and this will take a longer and greater effort.
In the end, it all comes down to how far you want to go down the road. Coding is not just putting together some instructions that work. It is so much more!
Good code is short, fast, elegant, easy to read and understand, simple, easy to modify and extend, easy to scale and refactor, and easy to test. It takes time to be able to write code that has all these qualities at the same time, but the good news is that you're taking the first step towards it at this very moment by reading this book. And I have no doubt you can do it. Anyone can, in fact, we all program all the time, only we aren't aware of it.
Would you like an example?
Say you want to make instant coffee. You have to get a mug, the instant coffee jar, a teaspoon, water, and the kettle. Even if you're not aware of it, you're evaluating a lot of data. You're making sure that there is water in the kettle as well as the kettle is plugged-in, that the mug is clean, and that there is enough coffee in the jar. Then, you boil the water and maybe in the meantime you put some coffee in the mug. When the water is ready, you pour it into the cup, and stir.
So, how is this programming?
Well, we gathered resources (the kettle, coffee, water, teaspoon, and mug) and we verified some conditions on them (kettle is plugged-in, mug is clean, there is enough coffee). Then we started two actions (boiling the water and putting coffee in the mug), and when both of them were completed, we finally ended the procedure by pouring water in the mug and stirring.
Can you see it? I have just described the high-level functionality of a coffee program. It wasn't that hard because this is what the brain does all day long: evaluate conditions, decide to take actions, carry out tasks, repeat some of them, and stop at some point. Clean objects, put them back, and so on.
All you need now is to learn how to deconstruct all those actions you do automatically in real life so that a computer can actually make some sense of them. And you need to learn a language as well, to instruct it.
So this is what this book is for. I'll tell you how to do it and I'll try to do that by means of many simple but focused examples (my favorite kind).
I love to make references to the real world when I teach coding; I believe they help people retain the concepts better. However, now is the time to be a bit more rigorous and see what coding is from a more technical perspective.
When we write code, we're instructing a computer on what are the things it has to do. Where does the action happen? In many places: the computer memory, hard drives, network cables, CPU, and so on. It's a whole "world", which most of the time is the representation of a subset of the real world.
If you write a piece of software that allows people to buy clothes online, you will have to represent real people, real clothes, real brands, sizes, and so on and so forth, within the boundaries of a program.
In order to do so, you will need to create and handle objects in the program you're writing. A person can be an object. A car is an object. A pair of socks is an object. Luckily, Python understands objects very well.
The two main features any object has are properties and methods. Let's take a person object as an example. Typically in a computer program, you'll represent people as customers or employees. The properties that you store against them are things like the name, the SSN, the age, if they have a driving license, their e-mail, gender, and so on. In a computer program, you store all the data you need in order to use an object for the purpose you're serving. If you are coding a website to sell clothes, you probably want to store the height and weight as well as other measures of your customers so that you can suggest the appropriate clothes for them. So, properties are characteristics of an object. We use them all the time: "Could you pass me that pen?" – "Which one?" – "The black one." Here, we used the "black" property of a pen to identify it (most likely amongst a blue and a red one).
Methods are things that an object can do. As a person, I have methods such as speak, walk, sleep, wake-up, eat, dream, write, read, and so on. All the things that I can do could be seen as methods of the objects that represents me.
So, now that you know what objects are and that they expose methods that you can run and properties that you can inspect, you're ready to start coding. Coding in fact is simply about managing those objects that live in the subset of the world that we're reproducing in our software. You can create, use, reuse, and delete objects as you please.
According to the Data Model chapter on the official Python documentation:
"Objects are Python's abstraction for data. All data in a Python program is represented by objects or by relations between objects."
We'll take a closer look at Python objects in Chapter 6, Advanced Concepts – OOP, Decorators, and Iterators. For now, all we need to know is that every object in Python has an ID (or identity), a type, and a value.
Once created, the identity of an object is never changed. It's a unique identifier for it, and it's used behind the scenes by Python to retrieve the object when we want to use it.
The type as well, never changes. The type tells what operations are supported by the object and the possible values that can be assigned to it.
We'll see Python's most important data types in Chapter 2, Built-in Data Types.
The value can either change or not. If it can, the object is said to be mutable, while when it cannot, the object is said to be immutable.
How do we use an object? We give it a name of course! When you give an object a name, then you can use the name to retrieve the object and use it.
In a more generic sense, objects such as numbers, strings (text), collections, and so on are associated with a name. Usually, we say that this name is the name of a variable. You can see the variable as being like a box, which you can use to hold data.
So, you have all the objects you need: what now? Well, we need to use them, right? We may want to send them over a network connection or store them in a database. Maybe display them on a web page or write them into a file. In order to do so, we need to react to a user filling in a form, or pressing a button, or opening a web page and performing a search. We react by running our code, evaluating conditions to choose which parts to execute, how many times, and under which circumstances.
And to do all this, basically we need a language. That's what Python is for. Python is the language we'll use together throughout this book to instruct the computer to do something for us.
Now, enough of this theoretical stuff, let's get started.
Python is the marvelous creature of Guido Van Rossum, a Dutch computer scientist and mathematician who decided to gift the world with a project he was playing around with over Christmas 1989. The language appeared to the public somewhere around 1991, and since then has evolved to be one of the leading programming languages used worldwide today.
I started programming when I was 7 years old, on a Commodore VIC 20, which was later replaced by its bigger brother, the Commodore 64. The language was BASIC. Later on, I landed on Pascal, Assembly, C, C++, Java, JavaScript, Visual Basic, PHP, ASP, ASP .NET, C#, and other minor languages I cannot even remember, but only when I landed on Python, I finally had that feeling that you have when you find the right couch in the shop. When all of your body parts are yelling, "Buy this one! This one is perfect for us!"
It took me about a day to get used to it. Its syntax is a bit different from what I was used to, and in general, I very rarely worked with a language that defines scoping with indentation. But after getting past that initial feeling of discomfort (like having new shoes), I just fell in love with it. Deeply. Let's see why.
Before we get into the gory details, let's get a sense of why someone would want to use Python (I would recommend you to read the Python page on Wikipedia to get a more detailed introduction).
To my mind, Python exposes the following qualities.
Python runs everywhere, and porting a program from Linux to Windows or Mac is usually just a matter of fixing paths and settings. Python is designed for portability and it takes care of operating system (OS) specific quirks behind interfaces that shield you from the pain of having to write code tailored to a specific platform.
Python is extremely logical and coherent. You can see it was designed by a brilliant computer scientist. Most of the time you can just guess how a method is called, if you don't know it.
You may not realize how important this is right now, especially if you are at the beginning, but this is a major feature. It means less cluttering in your head, less skimming through the documentation, and less need for mapping in your brain when you code.
According to Mark Lutz (Learning Python, 5th Edition, O'Reilly Media), a Python program is typically one-fifth to one-third the size of equivalent Java or C++ code. This means the job gets done faster. And faster is good. Faster means a faster response on the market. Less code not only means less code to write, but also less code to read (and professional coders read much more than they write), less code to maintain, to debug, and to refactor.
Another important aspect is that Python runs without the need of lengthy and time consuming compilation and linkage steps, so you don't have to wait to see the results of your work.
Python has an incredibly wide standard library (it's said to come with "batteries included"). If that wasn't enough, the Python community all over the world maintains a body of third party libraries, tailored to specific needs, which you can access freely at the Python Package Index (PyPI). When you code Python and you realize that you need a certain feature, in most cases, there is at least one library where that feature has already been implemented for you.
Python is heavily focused on readability, coherence, and quality. The language uniformity allows for high readability and this is crucial nowadays where code is more of a collective effort than a solo experience. Another important aspect of Python is its intrinsic multi-paradigm nature. You can use it as scripting language, but you also can exploit object-oriented, imperative, and functional programming styles. It is versatile.
Another important aspect is that Python can be extended and integrated with many other languages, which means that even when a company is using a different language as their mainstream tool, Python can come in and act as a glue agent between complex applications that need to talk to each other in some way. This is kind of an advanced topic, but in the real world, this feature is very important.
Last but not least, the fun of it! Working with Python is fun. I can code for 8 hours and leave the office happy and satisfied, alien to the struggle other coders have to endure because they use languages that don't provide them with the same amount of well-designed data structures and constructs. Python makes coding fun, no doubt about it. And fun promotes motivation and productivity.
These are the major aspects why I would recommend Python to everyone for. Of course, there are many other technical and advanced features that I could have talked about, but they don't really pertain to an introductory section like this one. They will come up naturally, chapter after chapter, in this book.
Portability
Python runs everywhere, and porting a program from Linux to Windows or Mac is usually just a matter of fixing paths and settings. Python is designed for portability and it takes care of operating system (OS) specific quirks behind interfaces that shield you from the pain of having to write code tailored to a specific platform.
Python is extremely logical and coherent. You can see it was designed by a brilliant computer scientist. Most of the time you can just guess how a method is called, if you don't know it.
You may not realize how important this is right now, especially if you are at the beginning, but this is a major feature. It means less cluttering in your head, less skimming through the documentation, and less need for mapping in your brain when you code.
According to Mark Lutz (Learning Python, 5th Edition, O'Reilly Media), a Python program is typically one-fifth to one-third the size of equivalent Java or C++ code. This means the job gets done faster. And faster is good. Faster means a faster response on the market. Less code not only means less code to write, but also less code to read (and professional coders read much more than they write), less code to maintain, to debug, and to refactor.
Another important aspect is that Python runs without the need of lengthy and time consuming compilation and linkage steps, so you don't have to wait to see the results of your work.
Python has an incredibly wide standard library (it's said to come with "batteries included"). If that wasn't enough, the Python community all over the world maintains a body of third party libraries, tailored to specific needs, which you can access freely at the Python Package Index (PyPI). When you code Python and you realize that you need a certain feature, in most cases, there is at least one library where that feature has already been implemented for you.
Python is heavily focused on readability, coherence, and quality. The language uniformity allows for high readability and this is crucial nowadays where code is more of a collective effort than a solo experience. Another important aspect of Python is its intrinsic multi-paradigm nature. You can use it as scripting language, but you also can exploit object-oriented, imperative, and functional programming styles. It is versatile.
Another important aspect is that Python can be extended and integrated with many other languages, which means that even when a company is using a different language as their mainstream tool, Python can come in and act as a glue agent between complex applications that need to talk to each other in some way. This is kind of an advanced topic, but in the real world, this feature is very important.
Last but not least, the fun of it! Working with Python is fun. I can code for 8 hours and leave the office happy and satisfied, alien to the struggle other coders have to endure because they use languages that don't provide them with the same amount of well-designed data structures and constructs. Python makes coding fun, no doubt about it. And fun promotes motivation and productivity.
These are the major aspects why I would recommend Python to everyone for. Of course, there are many other technical and advanced features that I could have talked about, but they don't really pertain to an introductory section like this one. They will come up naturally, chapter after chapter, in this book.
Coherence
Python is extremely logical and coherent. You can see it was designed by a brilliant computer scientist. Most of the time you can just guess how a method is called, if you don't know it.
You may not realize how important this is right now, especially if you are at the beginning, but this is a major feature. It means less cluttering in your head, less skimming through the documentation, and less need for mapping in your brain when you code.
According to Mark Lutz (Learning Python, 5th Edition, O'Reilly Media), a Python program is typically one-fifth to one-third the size of equivalent Java or C++ code. This means the job gets done faster. And faster is good. Faster means a faster response on the market. Less code not only means less code to write, but also less code to read (and professional coders read much more than they write), less code to maintain, to debug, and to refactor.
Another important aspect is that Python runs without the need of lengthy and time consuming compilation and linkage steps, so you don't have to wait to see the results of your work.
Python has an incredibly wide standard library (it's said to come with "batteries included"). If that wasn't enough, the Python community all over the world maintains a body of third party libraries, tailored to specific needs, which you can access freely at the Python Package Index (PyPI). When you code Python and you realize that you need a certain feature, in most cases, there is at least one library where that feature has already been implemented for you.
Python is heavily focused on readability, coherence, and quality. The language uniformity allows for high readability and this is crucial nowadays where code is more of a collective effort than a solo experience. Another important aspect of Python is its intrinsic multi-paradigm nature. You can use it as scripting language, but you also can exploit object-oriented, imperative, and functional programming styles. It is versatile.
Another important aspect is that Python can be extended and integrated with many other languages, which means that even when a company is using a different language as their mainstream tool, Python can come in and act as a glue agent between complex applications that need to talk to each other in some way. This is kind of an advanced topic, but in the real world, this feature is very important.
Last but not least, the fun of it! Working with Python is fun. I can code for 8 hours and leave the office happy and satisfied, alien to the struggle other coders have to endure because they use languages that don't provide them with the same amount of well-designed data structures and constructs. Python makes coding fun, no doubt about it. And fun promotes motivation and productivity.
These are the major aspects why I would recommend Python to everyone for. Of course, there are many other technical and advanced features that I could have talked about, but they don't really pertain to an introductory section like this one. They will come up naturally, chapter after chapter, in this book.
Developer productivity
According to Mark Lutz (Learning Python, 5th Edition, O'Reilly Media), a Python program is typically one-fifth to one-third the size of equivalent Java or C++ code. This means the job gets done faster. And faster is good. Faster means a faster response on the market. Less code not only means less code to write, but also less code to read (and professional coders read much more than they write), less code to maintain, to debug, and to refactor.
Another important aspect is that Python runs without the need of lengthy and time consuming compilation and linkage steps, so you don't have to wait to see the results of your work.
Python has an incredibly wide standard library (it's said to come with "batteries included"). If that wasn't enough, the Python community all over the world maintains a body of third party libraries, tailored to specific needs, which you can access freely at the Python Package Index (PyPI). When you code Python and you realize that you need a certain feature, in most cases, there is at least one library where that feature has already been implemented for you.
Python is heavily focused on readability, coherence, and quality. The language uniformity allows for high readability and this is crucial nowadays where code is more of a collective effort than a solo experience. Another important aspect of Python is its intrinsic multi-paradigm nature. You can use it as scripting language, but you also can exploit object-oriented, imperative, and functional programming styles. It is versatile.
Another important aspect is that Python can be extended and integrated with many other languages, which means that even when a company is using a different language as their mainstream tool, Python can come in and act as a glue agent between complex applications that need to talk to each other in some way. This is kind of an advanced topic, but in the real world, this feature is very important.
Last but not least, the fun of it! Working with Python is fun. I can code for 8 hours and leave the office happy and satisfied, alien to the struggle other coders have to endure because they use languages that don't provide them with the same amount of well-designed data structures and constructs. Python makes coding fun, no doubt about it. And fun promotes motivation and productivity.
These are the major aspects why I would recommend Python to everyone for. Of course, there are many other technical and advanced features that I could have talked about, but they don't really pertain to an introductory section like this one. They will come up naturally, chapter after chapter, in this book.
An extensive library
Python has an incredibly wide standard library (it's said to come with "batteries included"). If that wasn't enough, the Python community all over the world maintains a body of third party libraries, tailored to specific needs, which you can access freely at the Python Package Index (PyPI). When you code Python and you realize that you need a certain feature, in most cases, there is at least one library where that feature has already been implemented for you.
Python is heavily focused on readability, coherence, and quality. The language uniformity allows for high readability and this is crucial nowadays where code is more of a collective effort than a solo experience. Another important aspect of Python is its intrinsic multi-paradigm nature. You can use it as scripting language, but you also can exploit object-oriented, imperative, and functional programming styles. It is versatile.
Another important aspect is that Python can be extended and integrated with many other languages, which means that even when a company is using a different language as their mainstream tool, Python can come in and act as a glue agent between complex applications that need to talk to each other in some way. This is kind of an advanced topic, but in the real world, this feature is very important.
Last but not least, the fun of it! Working with Python is fun. I can code for 8 hours and leave the office happy and satisfied, alien to the struggle other coders have to endure because they use languages that don't provide them with the same amount of well-designed data structures and constructs. Python makes coding fun, no doubt about it. And fun promotes motivation and productivity.
These are the major aspects why I would recommend Python to everyone for. Of course, there are many other technical and advanced features that I could have talked about, but they don't really pertain to an introductory section like this one. They will come up naturally, chapter after chapter, in this book.
Software quality
Python is heavily focused on readability, coherence, and quality. The language uniformity allows for high readability and this is crucial nowadays where code is more of a collective effort than a solo experience. Another important aspect of Python is its intrinsic multi-paradigm nature. You can use it as scripting language, but you also can exploit object-oriented, imperative, and functional programming styles. It is versatile.
Another important aspect is that Python can be extended and integrated with many other languages, which means that even when a company is using a different language as their mainstream tool, Python can come in and act as a glue agent between complex applications that need to talk to each other in some way. This is kind of an advanced topic, but in the real world, this feature is very important.
Last but not least, the fun of it! Working with Python is fun. I can code for 8 hours and leave the office happy and satisfied, alien to the struggle other coders have to endure because they use languages that don't provide them with the same amount of well-designed data structures and constructs. Python makes coding fun, no doubt about it. And fun promotes motivation and productivity.
These are the major aspects why I would recommend Python to everyone for. Of course, there are many other technical and advanced features that I could have talked about, but they don't really pertain to an introductory section like this one. They will come up naturally, chapter after chapter, in this book.
Software integration
Another important aspect is that Python can be extended and integrated with many other languages, which means that even when a company is using a different language as their mainstream tool, Python can come in and act as a glue agent between complex applications that need to talk to each other in some way. This is kind of an advanced topic, but in the real world, this feature is very important.
Last but not least, the fun of it! Working with Python is fun. I can code for 8 hours and leave the office happy and satisfied, alien to the struggle other coders have to endure because they use languages that don't provide them with the same amount of well-designed data structures and constructs. Python makes coding fun, no doubt about it. And fun promotes motivation and productivity.
These are the major aspects why I would recommend Python to everyone for. Of course, there are many other technical and advanced features that I could have talked about, but they don't really pertain to an introductory section like this one. They will come up naturally, chapter after chapter, in this book.
Satisfaction and enjoyment
Last but not least, the fun of it! Working with Python is fun. I can code for 8 hours and leave the office happy and satisfied, alien to the struggle other coders have to endure because they use languages that don't provide them with the same amount of well-designed data structures and constructs. Python makes coding fun, no doubt about it. And fun promotes motivation and productivity.
These are the major aspects why I would recommend Python to everyone for. Of course, there are many other technical and advanced features that I could have talked about, but they don't really pertain to an introductory section like this one. They will come up naturally, chapter after chapter, in this book.
Probably, the only drawback that one could find in Python, which is not due to personal preferences, is the execution speed. Typically, Python is slower than its compiled brothers. The standard implementation of Python produces, when you run an application, a compiled version of the source code called byte code (with the extension .pyc
), which is then run by the Python interpreter. The advantage of this approach is portability, which we pay for with a slowdown due to the fact that Python is not compiled down to machine level as are other languages.
However, Python speed is rarely a problem today, hence its wide use regardless of this suboptimal feature. What happens is that in real life, hardware cost is no longer a problem, and usually it's easy enough to gain speed by parallelizing tasks. When it comes to number crunching though, one can switch to faster Python implementations, such as PyPy, which provides an average 7-fold speedup by implementing advanced compilation techniques (check http://pypy.org/ for reference).
When doing data science, you'll most likely find that the libraries that you use with Python, such as Pandas and Numpy, achieve native speed due to the way they are implemented.
If that wasn't a good enough argument, you can always consider that Python is driving the backend of services such as Spotify and Instagram, where performance is a concern. Nonetheless, Python does its job perfectly adequately.
Not yet convinced? Let's take a very brief look at the companies that are using Python today: Google, YouTube, Dropbox, Yahoo, Zope Corporation, Industrial Light & Magic, Walt Disney Feature Animation, Pixar, NASA, NSA, Red Hat, Nokia, IBM, Netflix, Yelp, Intel, Cisco, HP, Qualcomm, and JPMorgan Chase, just to name a few.
Even games such as Battlefield 2, Civilization 4, and QuArK are implemented using Python.
Python is used in many different contexts, such as system programming, web programming, GUI applications, gaming and robotics, rapid prototyping, system integration, data science, database applications, and much more.
Before we talk about installing Python on your system, let me tell you about which Python version I'll be using in this book.
Python comes in two main versions—Python 2, which is the past—and Python 3, which is the present. The two versions, though very similar, are incompatible on some aspects.
In the real world, Python 2 is actually quite far from being the past. In short, even though Python 3 has been out since 2008, the transition phase is still far from being over. This is mostly due to the fact that Python 2 is widely used in the industry, and of course, companies aren't so keen on updating their systems just for the sake of updating, following the if it ain't broke, don't fix it philosophy. You can read all about the transition between the two versions on the Web.
Another issue that was hindering the transition is the availability of third-party libraries. Usually, a Python project relies on tens of external libraries, and of course, when you start a new project, you need to be sure that there is already a version 3 compatible library for any business requirement that may come up. If that's not the case, starting a brand new project in Python 3 means introducing a potential risk, which many companies are not happy to take.
At the time of writing, the majority of the most widely used libraries have been ported to Python 3, and it's quite safe to start a project in Python 3 for most cases. Many of the libraries have been rewritten so that they are compatible with both versions, mostly harnessing the power of the six (2 x 3) library, which helps introspecting and adapting the behavior according to the version used.
On my Linux box (Ubuntu 14.04), I have the following Python version:
>>> import sys >>> print(sys.version) 3.4.0 (default, Apr 11 2014, 13:05:11) [GCC 4.8.2]
So you can see that my Python version is 3.4.0. The preceding text is a little bit of Python code that I typed into my console. We'll talk about it in a moment.
All the examples in this book will be run using this Python version. Most of them will run also in Python 2 (I have version 2.7.6 installed as well), and those that won't will just require some minor adjustments to cater for the small incompatibilities between the two versions. Another reason behind this choice is that I think it's better to learn Python 3, and then, if you need to, learn the differences it has with Python 2, rather than going the other way around.
Don't worry about this version thing though: it's not that big an issue in practice.
Python 2 versus Python 3 – the great debate
Python comes in two main versions—Python 2, which is the past—and Python 3, which is the present. The two versions, though very similar, are incompatible on some aspects.
In the real world, Python 2 is actually quite far from being the past. In short, even though Python 3 has been out since 2008, the transition phase is still far from being over. This is mostly due to the fact that Python 2 is widely used in the industry, and of course, companies aren't so keen on updating their systems just for the sake of updating, following the if it ain't broke, don't fix it philosophy. You can read all about the transition between the two versions on the Web.
Another issue that was hindering the transition is the availability of third-party libraries. Usually, a Python project relies on tens of external libraries, and of course, when you start a new project, you need to be sure that there is already a version 3 compatible library for any business requirement that may come up. If that's not the case, starting a brand new project in Python 3 means introducing a potential risk, which many companies are not happy to take.
At the time of writing, the majority of the most widely used libraries have been ported to Python 3, and it's quite safe to start a project in Python 3 for most cases. Many of the libraries have been rewritten so that they are compatible with both versions, mostly harnessing the power of the six (2 x 3) library, which helps introspecting and adapting the behavior according to the version used.
On my Linux box (Ubuntu 14.04), I have the following Python version:
>>> import sys >>> print(sys.version) 3.4.0 (default, Apr 11 2014, 13:05:11) [GCC 4.8.2]
So you can see that my Python version is 3.4.0. The preceding text is a little bit of Python code that I typed into my console. We'll talk about it in a moment.
All the examples in this book will be run using this Python version. Most of them will run also in Python 2 (I have version 2.7.6 installed as well), and those that won't will just require some minor adjustments to cater for the small incompatibilities between the two versions. Another reason behind this choice is that I think it's better to learn Python 3, and then, if you need to, learn the differences it has with Python 2, rather than going the other way around.
Don't worry about this version thing though: it's not that big an issue in practice.
I never really got the point of having a setup section in a book, regardless of what it is that you have to set up. Most of the time, between the time the author writes the instruction and the time you actually try them out, months have passed. That is, if you're lucky. One version change and things may not work the way it is described in the book. Luckily, we have the Web now, so in order to help you get up and running, I'll just give you pointers and objectives.
First of all, let's talk about your OS. Python is fully integrated and most likely already installed in basically almost every Linux distribution. If you have a Mac, it's likely that Python is already there as well (however, possibly only Python 2.7), whereas if you're using Windows, you probably need to install it.
Getting Python and the libraries you need up and running requires a bit of handiwork. Linux happens to be the most user friendly OS for Python programmers, Windows on the other hand is the one that requires the biggest effort, Mac being somewhere in between. For this reason, if you can choose, I suggest you to use Linux. If you can't, and you have a Mac, then go for it anyway. If you use Windows, you'll be fine for the examples in this book, but in general working with Python will require you a bit more tweaking.
My OS is Ubuntu 14.04, and this is what I will use throughout the book, along with Python 3.4.0.
The place you want to start is the official Python website: https://www.python.org. This website hosts the official Python documentation and many other resources that you will find very useful. Take the time to explore it.
Tip
Another excellent, resourceful website on Python and its ecosystem is http://docs.python-guide.org.
Find the download section and choose the installer for your OS. If you are on Windows, make sure that when you run the installer, you check the option install pip
(actually, I would suggest to make a complete installation, just to be safe, of all the components the installer holds). We'll talk about pip later.
Now that Python is installed in your system, the objective is to be able to open a console and run the Python interactive shell by typing python
.
To open the console in Windows, go to the Start menu, choose Run, and type cmd
. If you encounter anything that looks like a permission problem while working on the examples of this book, please make sure you are running the console with administrator rights.
On the Mac OS X, you can start a terminal by going to Applications | Utilities | Terminal.
If you are on Linux, you know all that there is to know about the console.
Whatever console you open, type python
at the prompt, and make sure the Python interactive shell shows up. Type exit()
to quit. Keep in mind that you may have to specify python3
if your OS comes with Python 2.* preinstalled.
This is how it should look on Windows 7:

And this is how it should look on Linux:

Now that Python is set up and you can run it, it's time to make sure you have the other tool that will be indispensable to follow the examples in the book: virtualenv.
As you probably have guessed by its name, virtualenv is all about virtual environments. Let me explain what they are and why we need them and let me do it by means of a simple example.
You install Python on your system and you start working on a website for client X. You create a project folder and start coding. Along the way you also install some libraries, for example the Django framework, which we'll see in depth in Chapter 10, Web Development Done Right. Let's say the Django version you install for project X is 1.7.1.
Now, your website is so good that you get another client, Y. He wants you to build another website, so you start project Y and, along the way, you need to install Django again. The only issue is that now the Django version is 1.8 and you cannot install it on your system because this would replace the version you installed for project X. You don't want to risk introducing incompatibility issues, so you have two choices: either you stick with the version you have currently on your machine, or you upgrade it and make sure the first project is still fully working correctly with the new version.
Let's be honest, neither of these options is very appealing, right? Definitely not. So, here's the solution: virtualenv!
virtualenv is a tool that allows you to create a virtual environment. In other words, it is a tool to create isolated Python environments, each of which is a folder that contains all the necessary executables to use the packages that a Python project would need (think of packages as libraries for the time being).
So you create a virtual environment for project X, install all the dependencies, and then you create a virtual environment for project Y, installing all its dependencies without the slightest worry because every library you install ends up within the boundaries of the appropriate virtual environment. In our example, project X will hold Django 1.7.1, while project Y will hold Django 1.8.
Note
It is of vital importance that you never install libraries directly at the system level. Linux for example relies on Python for many different tasks and operations, and if you fiddle with the system installation of Python, you risk compromising the integrity of the whole system (guess to whom this happened…). So take this as a rule, such as brushing your teeth before going to bed: always, always create a virtual environment when you start a new project.
To install virtualenv on your system, there are a few different ways. On a Debian-based distribution of Linux for example, you can install it with the following command:
$ sudo apt-get install python-virtualenv
Probably, the easiest way is to use pip
though, with the following command:
$ sudo pip install virtualenv # sudo may by optional
pip
is a package management system used to install and manage software packages written in Python.
Python 3 has built-in support for virtual environments, but in practice, the external libraries are still the default on production systems. If you have trouble getting virtualenv up and running, please refer to the virtualenv official website: https://virtualenv.pypa.io.
It is very easy to create a virtual environment, but according to how your system is configured and which Python version you want the virtual environment to run, you need to run the command properly. Another thing you will need to do with a virtualenv, when you want to work with it, is to activate it. Activating a virtualenv basically produces some path juggling behind the scenes so that when you call the Python interpreter, you're actually calling the active virtual environment one, instead of the mere system one.
I'll show you a full example on both Linux and Windows. We will:
- Create a folder named
learning.python
under your project root (which in my case is a folder calledsrv
, in my home folder). Please adapt the paths according to the setup you fancy on your box. - Within the
learning.python
folder, we will create a virtual environment called.lpvenv
.Note
Some developers prefer to call all virtual environments using the same name (for example,
.venv
). This way they can run scripts against any virtualenv by just knowing the name of the project they dwell in. This is a very common technique that I use as well. The dot in.venv
is because in Linux/Mac prepending a name with a dot makes that file or folder invisible. - After creating the virtual environment, we will activate it (this is slightly different between Linux, Mac, and Windows).
- Then, we'll make sure that we are running the desired Python version (3.4.*) by running the Python interactive shell.
- Finally, we will deactivate the virtual environment using the deactivate command.
These five simple steps will show you all you have to do to start and use a project.
Here's an example of how those steps might look like on Linux (commands that start with a #
are comments):

Notice that I had to explicitly tell virtualenv to use the Python 3.4 interpreter because on my box Python 2.7 is the default one. Had I not done that, I would have had a virtual environment with Python 2.7 instead of Python 3.4.
You can combine the two instructions for step 2 in one single command like this:
$ virtualenv -p $( which python3.4 ) .lpvenv
I preferred to be explicitly verbose in this instance, to help you understand each bit of the procedure.
Another thing to notice is that in order to activate a virtual environment, we need to run the /bin/activate
script, which needs to be sourced (when a script is "sourced", it means that its effects stick around when it's done running). This is very important. Also notice how the prompt changes after we activate the virtual environment, showing its name on the left (and how it disappears when we deactivate). In Mac OS, the steps are the same so I won't repeat them here.
Now let's have a look at how we can achieve the same result in Windows. You will probably have to play around a bit, especially if you have a different Windows or Python version than I'm using here. This is all good experience though, so try and think positively at the initial struggle that every coder has to go through in order to get things going.
Here's how it should look on Windows (commands that start with ::
are comments):

Notice there are a few small differences from the Linux version. Apart from the commands to create and navigate the folders, one important difference is how you activate your virtualenv. Also, in Windows there is no which
command, so we used the where
command.
At this point, you should be able to create and activate a virtual environment. Please try and create another one without me guiding you, get acquainted to this procedure because it's something that you will always be doing: we never work system-wide with Python, remember? It's extremely important.
So, with the scaffolding out of the way, we're ready to talk a bit more about Python and how you can use it. Before we do it though, allow me to spend a few words about the console.
In this era of GUIs and touchscreen devices, it seems a little ridiculous to have to resort to a tool such as the console, when everything is just about one click away.
But the truth is every time you remove your right hand from the keyboard (or the left one, if you're a lefty) to grab your mouse and move the cursor over to the spot you want to click, you're losing time. Getting things done with the console, counter-intuitively as it may be, results in higher productivity and speed. I know, you have to trust me on this.
Speed and productivity are important and personally, I have nothing against the mouse, but there is another very good reason for which you may want to get well acquainted with the console: when you develop code that ends up on some server, the console might be the only available tool. If you make friends with it, I promise you, you will never get lost when it's of utmost importance that you don't (typically, when the website is down and you have to investigate very quickly what's going on).
So it's really up to you. If you're in doubt, please grant me the benefit of the doubt and give it a try. It's easier than you think, and you'll never regret it. There is nothing more pitiful than a good developer who gets lost within an SSH connection to a server because they are used to their own custom set of tools, and only to that.
Now, let's get back to Python.
Setting up the Python interpreter
First of all, let's talk about your OS. Python is fully integrated and most likely already installed in basically almost every Linux distribution. If you have a Mac, it's likely that Python is already there as well (however, possibly only Python 2.7), whereas if you're using Windows, you probably need to install it.
Getting Python and the libraries you need up and running requires a bit of handiwork. Linux happens to be the most user friendly OS for Python programmers, Windows on the other hand is the one that requires the biggest effort, Mac being somewhere in between. For this reason, if you can choose, I suggest you to use Linux. If you can't, and you have a Mac, then go for it anyway. If you use Windows, you'll be fine for the examples in this book, but in general working with Python will require you a bit more tweaking.
My OS is Ubuntu 14.04, and this is what I will use throughout the book, along with Python 3.4.0.
The place you want to start is the official Python website: https://www.python.org. This website hosts the official Python documentation and many other resources that you will find very useful. Take the time to explore it.
Tip
Another excellent, resourceful website on Python and its ecosystem is http://docs.python-guide.org.
Find the download section and choose the installer for your OS. If you are on Windows, make sure that when you run the installer, you check the option install pip
(actually, I would suggest to make a complete installation, just to be safe, of all the components the installer holds). We'll talk about pip later.
Now that Python is installed in your system, the objective is to be able to open a console and run the Python interactive shell by typing python
.
To open the console in Windows, go to the Start menu, choose Run, and type cmd
. If you encounter anything that looks like a permission problem while working on the examples of this book, please make sure you are running the console with administrator rights.
On the Mac OS X, you can start a terminal by going to Applications | Utilities | Terminal.
If you are on Linux, you know all that there is to know about the console.
Whatever console you open, type python
at the prompt, and make sure the Python interactive shell shows up. Type exit()
to quit. Keep in mind that you may have to specify python3
if your OS comes with Python 2.* preinstalled.
This is how it should look on Windows 7:

And this is how it should look on Linux:

Now that Python is set up and you can run it, it's time to make sure you have the other tool that will be indispensable to follow the examples in the book: virtualenv.
As you probably have guessed by its name, virtualenv is all about virtual environments. Let me explain what they are and why we need them and let me do it by means of a simple example.
You install Python on your system and you start working on a website for client X. You create a project folder and start coding. Along the way you also install some libraries, for example the Django framework, which we'll see in depth in Chapter 10, Web Development Done Right. Let's say the Django version you install for project X is 1.7.1.
Now, your website is so good that you get another client, Y. He wants you to build another website, so you start project Y and, along the way, you need to install Django again. The only issue is that now the Django version is 1.8 and you cannot install it on your system because this would replace the version you installed for project X. You don't want to risk introducing incompatibility issues, so you have two choices: either you stick with the version you have currently on your machine, or you upgrade it and make sure the first project is still fully working correctly with the new version.
Let's be honest, neither of these options is very appealing, right? Definitely not. So, here's the solution: virtualenv!
virtualenv is a tool that allows you to create a virtual environment. In other words, it is a tool to create isolated Python environments, each of which is a folder that contains all the necessary executables to use the packages that a Python project would need (think of packages as libraries for the time being).
So you create a virtual environment for project X, install all the dependencies, and then you create a virtual environment for project Y, installing all its dependencies without the slightest worry because every library you install ends up within the boundaries of the appropriate virtual environment. In our example, project X will hold Django 1.7.1, while project Y will hold Django 1.8.
Note
It is of vital importance that you never install libraries directly at the system level. Linux for example relies on Python for many different tasks and operations, and if you fiddle with the system installation of Python, you risk compromising the integrity of the whole system (guess to whom this happened…). So take this as a rule, such as brushing your teeth before going to bed: always, always create a virtual environment when you start a new project.
To install virtualenv on your system, there are a few different ways. On a Debian-based distribution of Linux for example, you can install it with the following command:
$ sudo apt-get install python-virtualenv
Probably, the easiest way is to use pip
though, with the following command:
$ sudo pip install virtualenv # sudo may by optional
pip
is a package management system used to install and manage software packages written in Python.
Python 3 has built-in support for virtual environments, but in practice, the external libraries are still the default on production systems. If you have trouble getting virtualenv up and running, please refer to the virtualenv official website: https://virtualenv.pypa.io.
It is very easy to create a virtual environment, but according to how your system is configured and which Python version you want the virtual environment to run, you need to run the command properly. Another thing you will need to do with a virtualenv, when you want to work with it, is to activate it. Activating a virtualenv basically produces some path juggling behind the scenes so that when you call the Python interpreter, you're actually calling the active virtual environment one, instead of the mere system one.
I'll show you a full example on both Linux and Windows. We will:
- Create a folder named
learning.python
under your project root (which in my case is a folder calledsrv
, in my home folder). Please adapt the paths according to the setup you fancy on your box. - Within the
learning.python
folder, we will create a virtual environment called.lpvenv
.Note
Some developers prefer to call all virtual environments using the same name (for example,
.venv
). This way they can run scripts against any virtualenv by just knowing the name of the project they dwell in. This is a very common technique that I use as well. The dot in.venv
is because in Linux/Mac prepending a name with a dot makes that file or folder invisible. - After creating the virtual environment, we will activate it (this is slightly different between Linux, Mac, and Windows).
- Then, we'll make sure that we are running the desired Python version (3.4.*) by running the Python interactive shell.
- Finally, we will deactivate the virtual environment using the deactivate command.
These five simple steps will show you all you have to do to start and use a project.
Here's an example of how those steps might look like on Linux (commands that start with a #
are comments):

Notice that I had to explicitly tell virtualenv to use the Python 3.4 interpreter because on my box Python 2.7 is the default one. Had I not done that, I would have had a virtual environment with Python 2.7 instead of Python 3.4.
You can combine the two instructions for step 2 in one single command like this:
$ virtualenv -p $( which python3.4 ) .lpvenv
I preferred to be explicitly verbose in this instance, to help you understand each bit of the procedure.
Another thing to notice is that in order to activate a virtual environment, we need to run the /bin/activate
script, which needs to be sourced (when a script is "sourced", it means that its effects stick around when it's done running). This is very important. Also notice how the prompt changes after we activate the virtual environment, showing its name on the left (and how it disappears when we deactivate). In Mac OS, the steps are the same so I won't repeat them here.
Now let's have a look at how we can achieve the same result in Windows. You will probably have to play around a bit, especially if you have a different Windows or Python version than I'm using here. This is all good experience though, so try and think positively at the initial struggle that every coder has to go through in order to get things going.
Here's how it should look on Windows (commands that start with ::
are comments):

Notice there are a few small differences from the Linux version. Apart from the commands to create and navigate the folders, one important difference is how you activate your virtualenv. Also, in Windows there is no which
command, so we used the where
command.
At this point, you should be able to create and activate a virtual environment. Please try and create another one without me guiding you, get acquainted to this procedure because it's something that you will always be doing: we never work system-wide with Python, remember? It's extremely important.
So, with the scaffolding out of the way, we're ready to talk a bit more about Python and how you can use it. Before we do it though, allow me to spend a few words about the console.
In this era of GUIs and touchscreen devices, it seems a little ridiculous to have to resort to a tool such as the console, when everything is just about one click away.
But the truth is every time you remove your right hand from the keyboard (or the left one, if you're a lefty) to grab your mouse and move the cursor over to the spot you want to click, you're losing time. Getting things done with the console, counter-intuitively as it may be, results in higher productivity and speed. I know, you have to trust me on this.
Speed and productivity are important and personally, I have nothing against the mouse, but there is another very good reason for which you may want to get well acquainted with the console: when you develop code that ends up on some server, the console might be the only available tool. If you make friends with it, I promise you, you will never get lost when it's of utmost importance that you don't (typically, when the website is down and you have to investigate very quickly what's going on).
So it's really up to you. If you're in doubt, please grant me the benefit of the doubt and give it a try. It's easier than you think, and you'll never regret it. There is nothing more pitiful than a good developer who gets lost within an SSH connection to a server because they are used to their own custom set of tools, and only to that.
Now, let's get back to Python.
About virtualenv
As you probably have guessed by its name, virtualenv is all about virtual environments. Let me explain what they are and why we need them and let me do it by means of a simple example.
You install Python on your system and you start working on a website for client X. You create a project folder and start coding. Along the way you also install some libraries, for example the Django framework, which we'll see in depth in Chapter 10, Web Development Done Right. Let's say the Django version you install for project X is 1.7.1.
Now, your website is so good that you get another client, Y. He wants you to build another website, so you start project Y and, along the way, you need to install Django again. The only issue is that now the Django version is 1.8 and you cannot install it on your system because this would replace the version you installed for project X. You don't want to risk introducing incompatibility issues, so you have two choices: either you stick with the version you have currently on your machine, or you upgrade it and make sure the first project is still fully working correctly with the new version.
Let's be honest, neither of these options is very appealing, right? Definitely not. So, here's the solution: virtualenv!
virtualenv is a tool that allows you to create a virtual environment. In other words, it is a tool to create isolated Python environments, each of which is a folder that contains all the necessary executables to use the packages that a Python project would need (think of packages as libraries for the time being).
So you create a virtual environment for project X, install all the dependencies, and then you create a virtual environment for project Y, installing all its dependencies without the slightest worry because every library you install ends up within the boundaries of the appropriate virtual environment. In our example, project X will hold Django 1.7.1, while project Y will hold Django 1.8.
Note
It is of vital importance that you never install libraries directly at the system level. Linux for example relies on Python for many different tasks and operations, and if you fiddle with the system installation of Python, you risk compromising the integrity of the whole system (guess to whom this happened…). So take this as a rule, such as brushing your teeth before going to bed: always, always create a virtual environment when you start a new project.
To install virtualenv on your system, there are a few different ways. On a Debian-based distribution of Linux for example, you can install it with the following command:
$ sudo apt-get install python-virtualenv
Probably, the easiest way is to use pip
though, with the following command:
$ sudo pip install virtualenv # sudo may by optional
pip
is a package management system used to install and manage software packages written in Python.
Python 3 has built-in support for virtual environments, but in practice, the external libraries are still the default on production systems. If you have trouble getting virtualenv up and running, please refer to the virtualenv official website: https://virtualenv.pypa.io.
It is very easy to create a virtual environment, but according to how your system is configured and which Python version you want the virtual environment to run, you need to run the command properly. Another thing you will need to do with a virtualenv, when you want to work with it, is to activate it. Activating a virtualenv basically produces some path juggling behind the scenes so that when you call the Python interpreter, you're actually calling the active virtual environment one, instead of the mere system one.
I'll show you a full example on both Linux and Windows. We will:
- Create a folder named
learning.python
under your project root (which in my case is a folder calledsrv
, in my home folder). Please adapt the paths according to the setup you fancy on your box. - Within the
learning.python
folder, we will create a virtual environment called.lpvenv
.Note
Some developers prefer to call all virtual environments using the same name (for example,
.venv
). This way they can run scripts against any virtualenv by just knowing the name of the project they dwell in. This is a very common technique that I use as well. The dot in.venv
is because in Linux/Mac prepending a name with a dot makes that file or folder invisible. - After creating the virtual environment, we will activate it (this is slightly different between Linux, Mac, and Windows).
- Then, we'll make sure that we are running the desired Python version (3.4.*) by running the Python interactive shell.
- Finally, we will deactivate the virtual environment using the deactivate command.
These five simple steps will show you all you have to do to start and use a project.
Here's an example of how those steps might look like on Linux (commands that start with a #
are comments):

Notice that I had to explicitly tell virtualenv to use the Python 3.4 interpreter because on my box Python 2.7 is the default one. Had I not done that, I would have had a virtual environment with Python 2.7 instead of Python 3.4.
You can combine the two instructions for step 2 in one single command like this:
$ virtualenv -p $( which python3.4 ) .lpvenv
I preferred to be explicitly verbose in this instance, to help you understand each bit of the procedure.
Another thing to notice is that in order to activate a virtual environment, we need to run the /bin/activate
script, which needs to be sourced (when a script is "sourced", it means that its effects stick around when it's done running). This is very important. Also notice how the prompt changes after we activate the virtual environment, showing its name on the left (and how it disappears when we deactivate). In Mac OS, the steps are the same so I won't repeat them here.
Now let's have a look at how we can achieve the same result in Windows. You will probably have to play around a bit, especially if you have a different Windows or Python version than I'm using here. This is all good experience though, so try and think positively at the initial struggle that every coder has to go through in order to get things going.
Here's how it should look on Windows (commands that start with ::
are comments):

Notice there are a few small differences from the Linux version. Apart from the commands to create and navigate the folders, one important difference is how you activate your virtualenv. Also, in Windows there is no which
command, so we used the where
command.
At this point, you should be able to create and activate a virtual environment. Please try and create another one without me guiding you, get acquainted to this procedure because it's something that you will always be doing: we never work system-wide with Python, remember? It's extremely important.
So, with the scaffolding out of the way, we're ready to talk a bit more about Python and how you can use it. Before we do it though, allow me to spend a few words about the console.
In this era of GUIs and touchscreen devices, it seems a little ridiculous to have to resort to a tool such as the console, when everything is just about one click away.
But the truth is every time you remove your right hand from the keyboard (or the left one, if you're a lefty) to grab your mouse and move the cursor over to the spot you want to click, you're losing time. Getting things done with the console, counter-intuitively as it may be, results in higher productivity and speed. I know, you have to trust me on this.
Speed and productivity are important and personally, I have nothing against the mouse, but there is another very good reason for which you may want to get well acquainted with the console: when you develop code that ends up on some server, the console might be the only available tool. If you make friends with it, I promise you, you will never get lost when it's of utmost importance that you don't (typically, when the website is down and you have to investigate very quickly what's going on).
So it's really up to you. If you're in doubt, please grant me the benefit of the doubt and give it a try. It's easier than you think, and you'll never regret it. There is nothing more pitiful than a good developer who gets lost within an SSH connection to a server because they are used to their own custom set of tools, and only to that.
Now, let's get back to Python.
Your first virtual environment
It is very easy to create a virtual environment, but according to how your system is configured and which Python version you want the virtual environment to run, you need to run the command properly. Another thing you will need to do with a virtualenv, when you want to work with it, is to activate it. Activating a virtualenv basically produces some path juggling behind the scenes so that when you call the Python interpreter, you're actually calling the active virtual environment one, instead of the mere system one.
I'll show you a full example on both Linux and Windows. We will:
- Create a folder named
learning.python
under your project root (which in my case is a folder calledsrv
, in my home folder). Please adapt the paths according to the setup you fancy on your box. - Within the
learning.python
folder, we will create a virtual environment called.lpvenv
.Note
Some developers prefer to call all virtual environments using the same name (for example,
.venv
). This way they can run scripts against any virtualenv by just knowing the name of the project they dwell in. This is a very common technique that I use as well. The dot in.venv
is because in Linux/Mac prepending a name with a dot makes that file or folder invisible. - After creating the virtual environment, we will activate it (this is slightly different between Linux, Mac, and Windows).
- Then, we'll make sure that we are running the desired Python version (3.4.*) by running the Python interactive shell.
- Finally, we will deactivate the virtual environment using the deactivate command.
These five simple steps will show you all you have to do to start and use a project.
Here's an example of how those steps might look like on Linux (commands that start with a #
are comments):

Notice that I had to explicitly tell virtualenv to use the Python 3.4 interpreter because on my box Python 2.7 is the default one. Had I not done that, I would have had a virtual environment with Python 2.7 instead of Python 3.4.
You can combine the two instructions for step 2 in one single command like this:
$ virtualenv -p $( which python3.4 ) .lpvenv
I preferred to be explicitly verbose in this instance, to help you understand each bit of the procedure.
Another thing to notice is that in order to activate a virtual environment, we need to run the /bin/activate
script, which needs to be sourced (when a script is "sourced", it means that its effects stick around when it's done running). This is very important. Also notice how the prompt changes after we activate the virtual environment, showing its name on the left (and how it disappears when we deactivate). In Mac OS, the steps are the same so I won't repeat them here.
Now let's have a look at how we can achieve the same result in Windows. You will probably have to play around a bit, especially if you have a different Windows or Python version than I'm using here. This is all good experience though, so try and think positively at the initial struggle that every coder has to go through in order to get things going.
Here's how it should look on Windows (commands that start with ::
are comments):

Notice there are a few small differences from the Linux version. Apart from the commands to create and navigate the folders, one important difference is how you activate your virtualenv. Also, in Windows there is no which
command, so we used the where
command.
At this point, you should be able to create and activate a virtual environment. Please try and create another one without me guiding you, get acquainted to this procedure because it's something that you will always be doing: we never work system-wide with Python, remember? It's extremely important.
So, with the scaffolding out of the way, we're ready to talk a bit more about Python and how you can use it. Before we do it though, allow me to spend a few words about the console.
In this era of GUIs and touchscreen devices, it seems a little ridiculous to have to resort to a tool such as the console, when everything is just about one click away.
But the truth is every time you remove your right hand from the keyboard (or the left one, if you're a lefty) to grab your mouse and move the cursor over to the spot you want to click, you're losing time. Getting things done with the console, counter-intuitively as it may be, results in higher productivity and speed. I know, you have to trust me on this.
Speed and productivity are important and personally, I have nothing against the mouse, but there is another very good reason for which you may want to get well acquainted with the console: when you develop code that ends up on some server, the console might be the only available tool. If you make friends with it, I promise you, you will never get lost when it's of utmost importance that you don't (typically, when the website is down and you have to investigate very quickly what's going on).
So it's really up to you. If you're in doubt, please grant me the benefit of the doubt and give it a try. It's easier than you think, and you'll never regret it. There is nothing more pitiful than a good developer who gets lost within an SSH connection to a server because they are used to their own custom set of tools, and only to that.
Now, let's get back to Python.
Your friend, the console
In this era of GUIs and touchscreen devices, it seems a little ridiculous to have to resort to a tool such as the console, when everything is just about one click away.
But the truth is every time you remove your right hand from the keyboard (or the left one, if you're a lefty) to grab your mouse and move the cursor over to the spot you want to click, you're losing time. Getting things done with the console, counter-intuitively as it may be, results in higher productivity and speed. I know, you have to trust me on this.
Speed and productivity are important and personally, I have nothing against the mouse, but there is another very good reason for which you may want to get well acquainted with the console: when you develop code that ends up on some server, the console might be the only available tool. If you make friends with it, I promise you, you will never get lost when it's of utmost importance that you don't (typically, when the website is down and you have to investigate very quickly what's going on).
So it's really up to you. If you're in doubt, please grant me the benefit of the doubt and give it a try. It's easier than you think, and you'll never regret it. There is nothing more pitiful than a good developer who gets lost within an SSH connection to a server because they are used to their own custom set of tools, and only to that.
Now, let's get back to Python.
There are a few different ways in which you can run a Python program.
Python can be used as a scripting language. In fact, it always proves itself very useful. Scripts are files (usually of small dimensions) that you normally execute to do something like a task. Many developers end up having their own arsenal of tools that they fire when they need to perform a task. For example, you can have scripts to parse data in a format and render it into another different format. Or you can use a script to work with files and folders. You can create or modify configuration files, and much more. Technically, there is not much that cannot be done in a script.
It's quite common to have scripts running at a precise time on a server. For example, if your website database needs cleaning every 24 hours (for example, the table that stores the user sessions, which expire pretty quickly but aren't cleaned automatically), you could set up a cron job that fires your script at 3:00 A.M. every day.
I have Python scripts to do all the menial tasks that would take me minutes or more to do manually, and at some point, I decided to automate. For example, I have a laptop that doesn't have a Fn key to toggle the touchpad on and off. I find this very annoying, and I don't want to go clicking about through several menus when I need to do it, so I wrote a small script that is smart enough to tell my system to toggle the touchpad active state, and now I can do it with one simple click from my launcher. Priceless.
We'll devote half of Chapter 8, The Edges – GUIs and Scripts on scripting with Python.
Another way of running Python is by calling the interactive shell. This is something we already saw when we typed python
on the command line of our console.
So open a console, activate your virtual environment (which by now should be second nature to you, right?), and type python
. You will be presented with a couple of lines that should look like this (if you are on Linux):
Python 3.4.0 (default, Apr 11 2014, 13:05:11) [GCC 4.8.2] on linux Type "help", "copyright", "credits" or "license" for more information.
Those >>>
are the prompt of the shell. They tell you that Python is waiting for you to type something. If you type a simple instruction, something that fits in one line, that's all you'll see. However, if you type something that requires more than one line of code, the shell will change the prompt to ...
, giving you a visual clue that you're typing a multiline statement (or anything that would require more than one line of code).
Go on, try it out, let's do some basic maths:
>>> 2 + 4 6 >>> 10 / 4 2.5 >>> 2 ** 1024 179769313486231590772930519078902473361797697894230657273430081157732675805500963132708477322407536021120113879871393357658789768814416622492847430639474124377767893424865485276302219601246094119453082952085005768838150682342462881473913110540827237163350510684586298239947245938479716304835356329624224137216
The last operation is showing you something incredible. We raise 2 to the power of 1024, and Python is handling this task with no trouble at all. Try to do it in Java, C++, or C#. It won't work, unless you use special libraries to handle such big numbers.
I use the interactive shell every day. It's extremely useful to debug very quickly, for example, to check if a data structure supports an operation. Or maybe to inspect or run a piece of code.
When you use Django (a web framework), the interactive shell is coupled with it and allows you to work your way through the framework tools, to inspect the data in the database, and many more things. You will find that the interactive shell will soon become one of your dearest friends on the journey you are embarking on.
Another solution, which comes in a much nicer graphic layout, is to use IDLE (Integrated DeveLopment Environment). It's quite a simple IDE, which is intended mostly for beginners. It has a slightly larger set of capabilities than the naked interactive shell you get in the console, so you may want to explore it. It comes for free in the Windows Python installer and you can easily install it in any other system. You can find information about it on the Python website.
Guido Van Rossum named Python after the British comedy group Monty Python, so it's rumored that the name IDLE has been chosen in honor of Erik Idle, one of Monty Python's founding members.
Apart from being run as a script, and within the boundaries of a shell, Python can be coded and run as proper software. We'll see many examples throughout the book about this mode. And we'll understand more about it in a moment, when we'll talk about how Python code is organized and run.
Python can also be run as a GUI (Graphical User Interface). There are several frameworks available, some of which are cross-platform and some others are platform-specific. In Chapter 8, The Edges – GUIs and Scripts, we'll see an example of a GUI application created using Tkinter, which is an object-oriented layer that lives on top of Tk (Tkinter means Tk Interface).
Note
Tk is a graphical user interface toolkit that takes desktop application development to a higher level than the conventional approach. It is the standard GUI for Tcl (Tool Command Language), but also for many other dynamic languages and can produce rich native applications that run seamlessly under Windows, Linux, Mac OS X, and more.
Tkinter comes bundled with Python, therefore it gives the programmer easy access to the GUI world, and for these reasons, I have chosen it to be the framework for the GUI examples that I'll present in this book.
Among the other GUI frameworks, we find that the following are the most widely used:
- PyQt
- wxPython
- PyGtk
Describing them in detail is outside the scope of this book, but you can find all the information you need on the Python website in the GUI Programming section. If GUIs are what you're looking for, remember to choose the one you want according to some principles. Make sure they:
- Offer all the features you may need to develop your project
- Run on all the platforms you may need to support
- Rely on a community that is as wide and active as possible
- Wrap graphic drivers/tools that you can easily install/access
Running Python scripts
Python can be used as a scripting language. In fact, it always proves itself very useful. Scripts are files (usually of small dimensions) that you normally execute to do something like a task. Many developers end up having their own arsenal of tools that they fire when they need to perform a task. For example, you can have scripts to parse data in a format and render it into another different format. Or you can use a script to work with files and folders. You can create or modify configuration files, and much more. Technically, there is not much that cannot be done in a script.
It's quite common to have scripts running at a precise time on a server. For example, if your website database needs cleaning every 24 hours (for example, the table that stores the user sessions, which expire pretty quickly but aren't cleaned automatically), you could set up a cron job that fires your script at 3:00 A.M. every day.
I have Python scripts to do all the menial tasks that would take me minutes or more to do manually, and at some point, I decided to automate. For example, I have a laptop that doesn't have a Fn key to toggle the touchpad on and off. I find this very annoying, and I don't want to go clicking about through several menus when I need to do it, so I wrote a small script that is smart enough to tell my system to toggle the touchpad active state, and now I can do it with one simple click from my launcher. Priceless.
We'll devote half of Chapter 8, The Edges – GUIs and Scripts on scripting with Python.
Another way of running Python is by calling the interactive shell. This is something we already saw when we typed python
on the command line of our console.
So open a console, activate your virtual environment (which by now should be second nature to you, right?), and type python
. You will be presented with a couple of lines that should look like this (if you are on Linux):
Python 3.4.0 (default, Apr 11 2014, 13:05:11) [GCC 4.8.2] on linux Type "help", "copyright", "credits" or "license" for more information.
Those >>>
are the prompt of the shell. They tell you that Python is waiting for you to type something. If you type a simple instruction, something that fits in one line, that's all you'll see. However, if you type something that requires more than one line of code, the shell will change the prompt to ...
, giving you a visual clue that you're typing a multiline statement (or anything that would require more than one line of code).
Go on, try it out, let's do some basic maths:
>>> 2 + 4 6 >>> 10 / 4 2.5 >>> 2 ** 1024 179769313486231590772930519078902473361797697894230657273430081157732675805500963132708477322407536021120113879871393357658789768814416622492847430639474124377767893424865485276302219601246094119453082952085005768838150682342462881473913110540827237163350510684586298239947245938479716304835356329624224137216
The last operation is showing you something incredible. We raise 2 to the power of 1024, and Python is handling this task with no trouble at all. Try to do it in Java, C++, or C#. It won't work, unless you use special libraries to handle such big numbers.
I use the interactive shell every day. It's extremely useful to debug very quickly, for example, to check if a data structure supports an operation. Or maybe to inspect or run a piece of code.
When you use Django (a web framework), the interactive shell is coupled with it and allows you to work your way through the framework tools, to inspect the data in the database, and many more things. You will find that the interactive shell will soon become one of your dearest friends on the journey you are embarking on.
Another solution, which comes in a much nicer graphic layout, is to use IDLE (Integrated DeveLopment Environment). It's quite a simple IDE, which is intended mostly for beginners. It has a slightly larger set of capabilities than the naked interactive shell you get in the console, so you may want to explore it. It comes for free in the Windows Python installer and you can easily install it in any other system. You can find information about it on the Python website.
Guido Van Rossum named Python after the British comedy group Monty Python, so it's rumored that the name IDLE has been chosen in honor of Erik Idle, one of Monty Python's founding members.
Apart from being run as a script, and within the boundaries of a shell, Python can be coded and run as proper software. We'll see many examples throughout the book about this mode. And we'll understand more about it in a moment, when we'll talk about how Python code is organized and run.
Python can also be run as a GUI (Graphical User Interface). There are several frameworks available, some of which are cross-platform and some others are platform-specific. In Chapter 8, The Edges – GUIs and Scripts, we'll see an example of a GUI application created using Tkinter, which is an object-oriented layer that lives on top of Tk (Tkinter means Tk Interface).
Note
Tk is a graphical user interface toolkit that takes desktop application development to a higher level than the conventional approach. It is the standard GUI for Tcl (Tool Command Language), but also for many other dynamic languages and can produce rich native applications that run seamlessly under Windows, Linux, Mac OS X, and more.
Tkinter comes bundled with Python, therefore it gives the programmer easy access to the GUI world, and for these reasons, I have chosen it to be the framework for the GUI examples that I'll present in this book.
Among the other GUI frameworks, we find that the following are the most widely used:
- PyQt
- wxPython
- PyGtk
Describing them in detail is outside the scope of this book, but you can find all the information you need on the Python website in the GUI Programming section. If GUIs are what you're looking for, remember to choose the one you want according to some principles. Make sure they:
- Offer all the features you may need to develop your project
- Run on all the platforms you may need to support
- Rely on a community that is as wide and active as possible
- Wrap graphic drivers/tools that you can easily install/access
Running the Python interactive shell
Another way of running Python is by calling the interactive shell. This is something we already saw when we typed python
on the command line of our console.
So open a console, activate your virtual environment (which by now should be second nature to you, right?), and type python
. You will be presented with a couple of lines that should look like this (if you are on Linux):
Python 3.4.0 (default, Apr 11 2014, 13:05:11) [GCC 4.8.2] on linux Type "help", "copyright", "credits" or "license" for more information.
Those >>>
are the prompt of the shell. They tell you that Python is waiting for you to type something. If you type a simple instruction, something that fits in one line, that's all you'll see. However, if you type something that requires more than one line of code, the shell will change the prompt to ...
, giving you a visual clue that you're typing a multiline statement (or anything that would require more than one line of code).
Go on, try it out, let's do some basic maths:
>>> 2 + 4 6 >>> 10 / 4 2.5 >>> 2 ** 1024 179769313486231590772930519078902473361797697894230657273430081157732675805500963132708477322407536021120113879871393357658789768814416622492847430639474124377767893424865485276302219601246094119453082952085005768838150682342462881473913110540827237163350510684586298239947245938479716304835356329624224137216
The last operation is showing you something incredible. We raise 2 to the power of 1024, and Python is handling this task with no trouble at all. Try to do it in Java, C++, or C#. It won't work, unless you use special libraries to handle such big numbers.
I use the interactive shell every day. It's extremely useful to debug very quickly, for example, to check if a data structure supports an operation. Or maybe to inspect or run a piece of code.
When you use Django (a web framework), the interactive shell is coupled with it and allows you to work your way through the framework tools, to inspect the data in the database, and many more things. You will find that the interactive shell will soon become one of your dearest friends on the journey you are embarking on.
Another solution, which comes in a much nicer graphic layout, is to use IDLE (Integrated DeveLopment Environment). It's quite a simple IDE, which is intended mostly for beginners. It has a slightly larger set of capabilities than the naked interactive shell you get in the console, so you may want to explore it. It comes for free in the Windows Python installer and you can easily install it in any other system. You can find information about it on the Python website.
Guido Van Rossum named Python after the British comedy group Monty Python, so it's rumored that the name IDLE has been chosen in honor of Erik Idle, one of Monty Python's founding members.
Apart from being run as a script, and within the boundaries of a shell, Python can be coded and run as proper software. We'll see many examples throughout the book about this mode. And we'll understand more about it in a moment, when we'll talk about how Python code is organized and run.
Python can also be run as a GUI (Graphical User Interface). There are several frameworks available, some of which are cross-platform and some others are platform-specific. In Chapter 8, The Edges – GUIs and Scripts, we'll see an example of a GUI application created using Tkinter, which is an object-oriented layer that lives on top of Tk (Tkinter means Tk Interface).
Note
Tk is a graphical user interface toolkit that takes desktop application development to a higher level than the conventional approach. It is the standard GUI for Tcl (Tool Command Language), but also for many other dynamic languages and can produce rich native applications that run seamlessly under Windows, Linux, Mac OS X, and more.
Tkinter comes bundled with Python, therefore it gives the programmer easy access to the GUI world, and for these reasons, I have chosen it to be the framework for the GUI examples that I'll present in this book.
Among the other GUI frameworks, we find that the following are the most widely used:
- PyQt
- wxPython
- PyGtk
Describing them in detail is outside the scope of this book, but you can find all the information you need on the Python website in the GUI Programming section. If GUIs are what you're looking for, remember to choose the one you want according to some principles. Make sure they:
- Offer all the features you may need to develop your project
- Run on all the platforms you may need to support
- Rely on a community that is as wide and active as possible
- Wrap graphic drivers/tools that you can easily install/access
Running Python as a service
Apart from being run as a script, and within the boundaries of a shell, Python can be coded and run as proper software. We'll see many examples throughout the book about this mode. And we'll understand more about it in a moment, when we'll talk about how Python code is organized and run.
Python can also be run as a GUI (Graphical User Interface). There are several frameworks available, some of which are cross-platform and some others are platform-specific. In Chapter 8, The Edges – GUIs and Scripts, we'll see an example of a GUI application created using Tkinter, which is an object-oriented layer that lives on top of Tk (Tkinter means Tk Interface).
Note
Tk is a graphical user interface toolkit that takes desktop application development to a higher level than the conventional approach. It is the standard GUI for Tcl (Tool Command Language), but also for many other dynamic languages and can produce rich native applications that run seamlessly under Windows, Linux, Mac OS X, and more.
Tkinter comes bundled with Python, therefore it gives the programmer easy access to the GUI world, and for these reasons, I have chosen it to be the framework for the GUI examples that I'll present in this book.
Among the other GUI frameworks, we find that the following are the most widely used:
- PyQt
- wxPython
- PyGtk
Describing them in detail is outside the scope of this book, but you can find all the information you need on the Python website in the GUI Programming section. If GUIs are what you're looking for, remember to choose the one you want according to some principles. Make sure they:
- Offer all the features you may need to develop your project
- Run on all the platforms you may need to support
- Rely on a community that is as wide and active as possible
- Wrap graphic drivers/tools that you can easily install/access
Running Python as a GUI application
Python can also be run as a GUI (Graphical User Interface). There are several frameworks available, some of which are cross-platform and some others are platform-specific. In Chapter 8, The Edges – GUIs and Scripts, we'll see an example of a GUI application created using Tkinter, which is an object-oriented layer that lives on top of Tk (Tkinter means Tk Interface).
Note
Tk is a graphical user interface toolkit that takes desktop application development to a higher level than the conventional approach. It is the standard GUI for Tcl (Tool Command Language), but also for many other dynamic languages and can produce rich native applications that run seamlessly under Windows, Linux, Mac OS X, and more.
Tkinter comes bundled with Python, therefore it gives the programmer easy access to the GUI world, and for these reasons, I have chosen it to be the framework for the GUI examples that I'll present in this book.
Among the other GUI frameworks, we find that the following are the most widely used:
- PyQt
- wxPython
- PyGtk
Describing them in detail is outside the scope of this book, but you can find all the information you need on the Python website in the GUI Programming section. If GUIs are what you're looking for, remember to choose the one you want according to some principles. Make sure they:
- Offer all the features you may need to develop your project
- Run on all the platforms you may need to support
- Rely on a community that is as wide and active as possible
- Wrap graphic drivers/tools that you can easily install/access
Let's talk a little bit about how Python code is organized. In this paragraph, we'll start going down the rabbit hole a little bit more and introduce a bit more technical names and concepts.
Starting with the basics, how is Python code organized? Of course, you write your code into files. When you save a file with the extension .py
, that file is said to be a Python module.
It would be impractical to save all the code that it is required for software to work within one single file. That solution works for scripts, which are usually not longer than a few hundred lines (and often they are quite shorter than that).
A complete Python application can be made of hundreds of thousands of lines of code, so you will have to scatter it through different modules. Better, but not nearly good enough. It turns out that even like this it would still be impractical to work with the code. So Python gives you another structure, called package, which allows you to group modules together. A package is nothing more than a folder, which must contain a special file, __init__.py
that doesn't need to hold any code but whose presence is required to tell Python that the folder is not just some folder, but it's actually a package (note that as of Python 3.3 __init__.py
is not strictly required any more).
As always, an example will make all of this much clearer. I have created an example structure in my book project, and when I type in my Linux console:
$ tree -v example
I get a tree representation of the contents of the ch1/example
folder, which holds the code for the examples of this chapter. Here's how a structure of a real simple application could look like:
example/ ├── core.py ├── run.py └── util ├── __init__.py ├── db.py ├── math.py └── network.py
You can see that within the root of this example, we have two modules, core.py
and run.py
, and one package: util
. Within core.py
, there may be the core logic of our application. On the other hand, within the run.py
module, we can probably find the logic to start the application. Within the util
package, I expect to find various utility tools, and in fact, we can guess that the modules there are called by the type of tools they hold: db.py
would hold tools to work with databases, math.py
would of course hold mathematical tools (maybe our application deals with financial data), and network.py
would probably hold tools to send/receive data on networks.
As explained before, the __init__.py
file is there just to tell Python that util
is a package and not just a mere folder.
Had this software been organized within modules only, it would have been much harder to infer its structure. I put a module only example under the ch1/files_only
folder, see it for yourself:
$ tree -v files_only
This shows us a completely different picture:
files_only/ ├── core.py ├── db.py ├── math.py ├── network.py └── run.py
It is a little harder to guess what each module does, right? Now, consider that this is just a simple example, so you can guess how much harder it would be to understand a real application if we couldn't organize the code in packages and modules.
When a developer is writing an application, it is very likely that they will need to apply the same piece of logic in different parts of it. For example, when writing a parser for the data that comes from a form that a user can fill in a web page, the application will have to validate whether a certain field is holding a number or not. Regardless of how the logic for this kind of validation is written, it's very likely that it will be needed in more than one place. For example in a poll application, where the user is asked many question, it's likely that several of them will require a numeric answer. For example:
- What is your age
- How many pets do you own
- How many children do you have
- How many times have you been married
It would be very bad practice to copy paste (or, more properly said: duplicate) the validation logic in every place where we expect a numeric answer. This would violate the DRY (Don't Repeat Yourself) principle, which states that you should never repeat the same piece of code more than once in your application. I feel the need to stress the importance of this principle: you should never repeat the same piece of code more than once in your application (got the irony?).
There are several reasons why repeating the same piece of logic can be very bad, the most important ones being:
- There could be a bug in the logic, and therefore, you would have to correct it in every place that logic is applied.
- You may want to amend the way you carry out the validation, and again you would have to change it in every place it is applied.
- You may forget to fix/amend a piece of logic because you missed it when searching for all its occurrences. This would leave wrong/inconsistent behavior in your application.
- Your code would be longer than needed, for no good reason.
Python is a wonderful language and provides you with all the tools you need to apply all the coding best practices. For this particular example, we need to be able to reuse a piece of code. To be able to reuse a piece of code, we need to have a construct that will hold the code for us so that we can call that construct every time we need to repeat the logic inside it. That construct exists, and it's called function.
I'm not going too deep into the specifics here, so please just remember that a function is a block of organized, reusable code which is used to perform a task. Functions can assume many forms and names, according to what kind of environment they belong to, but for now this is not important. We'll see the details when we are able to appreciate them, later on, in the book. Functions are the building blocks of modularity in your application, and they are almost indispensable (unless you're writing a super simple script, you'll use functions all the time). We'll explore functions in Chapter 4, Functions, the Building Blocks of Code.
Python comes with a very extensive library, as I already said a few pages ago. Now, maybe it's a good time to define what a library is: a library is a collection of functions and objects that provide functionalities that enrich the abilities of a language.
For example, within Python's math
library we can find a plethora of functions, one of which is the factorial
function, which of course calculates the factorial of a number.
Note
In mathematics, the factorial of a non-negative integer number N, denoted as N!, is defined as the product of all positive integers less than or equal to N. For example, the factorial of 5 is calculated as:
5! = 5 * 4 * 3 * 2 * 1 = 120
The factorial of 0
is 0! = 1
, to respect the convention for an empty product.
So, if you wanted to use this function in your code, all you would have to do is to import it and call it with the right input values. Don't worry too much if input values and the concept of calling is not very clear for now, please just concentrate on the import part.
In Python, to calculate the factorial of number 5, we just need the following code:
>>> from math import factorial >>> factorial(5) 120
So, let's go back to our example, the one with core.py
, run.py
, util
, and so on.
In our example, the package util
is our utility library. Our custom utility belt that holds all those reusable tools (that is, functions), which we need in our application. Some of them will deal with databases (db.py
), some with the network (network.py
), and some will perform mathematical calculations (math.py
) that are outside the scope of Python's standard math
library and therefore, we had to code them for ourselves.
We will see in detail how to import functions and use them in their dedicated chapter. Let's now talk about another very important concept: Python's execution model.
How do we use modules and packages
When a developer is writing an application, it is very likely that they will need to apply the same piece of logic in different parts of it. For example, when writing a parser for the data that comes from a form that a user can fill in a web page, the application will have to validate whether a certain field is holding a number or not. Regardless of how the logic for this kind of validation is written, it's very likely that it will be needed in more than one place. For example in a poll application, where the user is asked many question, it's likely that several of them will require a numeric answer. For example:
- What is your age
- How many pets do you own
- How many children do you have
- How many times have you been married
It would be very bad practice to copy paste (or, more properly said: duplicate) the validation logic in every place where we expect a numeric answer. This would violate the DRY (Don't Repeat Yourself) principle, which states that you should never repeat the same piece of code more than once in your application. I feel the need to stress the importance of this principle: you should never repeat the same piece of code more than once in your application (got the irony?).
There are several reasons why repeating the same piece of logic can be very bad, the most important ones being:
- There could be a bug in the logic, and therefore, you would have to correct it in every place that logic is applied.
- You may want to amend the way you carry out the validation, and again you would have to change it in every place it is applied.
- You may forget to fix/amend a piece of logic because you missed it when searching for all its occurrences. This would leave wrong/inconsistent behavior in your application.
- Your code would be longer than needed, for no good reason.
Python is a wonderful language and provides you with all the tools you need to apply all the coding best practices. For this particular example, we need to be able to reuse a piece of code. To be able to reuse a piece of code, we need to have a construct that will hold the code for us so that we can call that construct every time we need to repeat the logic inside it. That construct exists, and it's called function.
I'm not going too deep into the specifics here, so please just remember that a function is a block of organized, reusable code which is used to perform a task. Functions can assume many forms and names, according to what kind of environment they belong to, but for now this is not important. We'll see the details when we are able to appreciate them, later on, in the book. Functions are the building blocks of modularity in your application, and they are almost indispensable (unless you're writing a super simple script, you'll use functions all the time). We'll explore functions in Chapter 4, Functions, the Building Blocks of Code.
Python comes with a very extensive library, as I already said a few pages ago. Now, maybe it's a good time to define what a library is: a library is a collection of functions and objects that provide functionalities that enrich the abilities of a language.
For example, within Python's math
library we can find a plethora of functions, one of which is the factorial
function, which of course calculates the factorial of a number.
Note
In mathematics, the factorial of a non-negative integer number N, denoted as N!, is defined as the product of all positive integers less than or equal to N. For example, the factorial of 5 is calculated as:
5! = 5 * 4 * 3 * 2 * 1 = 120
The factorial of 0
is 0! = 1
, to respect the convention for an empty product.
So, if you wanted to use this function in your code, all you would have to do is to import it and call it with the right input values. Don't worry too much if input values and the concept of calling is not very clear for now, please just concentrate on the import part.
In Python, to calculate the factorial of number 5, we just need the following code:
>>> from math import factorial >>> factorial(5) 120
So, let's go back to our example, the one with core.py
, run.py
, util
, and so on.
In our example, the package util
is our utility library. Our custom utility belt that holds all those reusable tools (that is, functions), which we need in our application. Some of them will deal with databases (db.py
), some with the network (network.py
), and some will perform mathematical calculations (math.py
) that are outside the scope of Python's standard math
library and therefore, we had to code them for ourselves.
We will see in detail how to import functions and use them in their dedicated chapter. Let's now talk about another very important concept: Python's execution model.
In this paragraph, I would like to introduce you to a few very important concepts, such as scope, names, and namespaces. You can read all about Python's execution model in the official Language reference, of course, but I would argue that it is quite technical and abstract, so let me give you a less formal explanation first.
Say you are looking for a book, so you go to the library and ask someone for the book you want to fetch. They tell you something like "second floor, section X, row three". So you go up the stairs, look for section X, and so on.
It would be very different to enter a library where all the books are piled together in random order in one big room. No floors, no sections, no rows, no order. Fetching a book would be extremely hard.
When we write code we have the same issue: we have to try and organize it so that it will be easy for someone who has no prior knowledge about it to find what they're looking for. When software is structured correctly, it also promotes code reuse. On the other hand, disorganized software is more likely to expose scattered pieces of duplicated logic.
First of all, let's start with the book. We refer to a book by its title and in Python lingo, that would be a name. Python names are the closest abstraction to what other languages call variables. Names basically refer to objects and are introduced by name binding operations. Let's make a quick example (notice that anything that follows a #
is a comment):
>>> n = 3 # integer number >>> address = "221b Baker Street, NW1 6XE, London" # S. Holmes >>> employee = { ... 'age': 45, ... 'role': 'CTO', ... 'SSN': 'AB1234567', ... } >>> # let's print them >>> n 3 >>> address '221b Baker Street, NW1 6XE, London' >>> employee {'role': 'CTO', 'SSN': 'AB1234567', 'age': 45} >>> # what if I try to print a name I didn't define? >>> other_name Traceback (most recent call last): File "<stdin>", line 1, in <module> NameError: name 'other_name' is not defined
We defined three objects in the preceding code (do you remember what are the three features every Python object has?):
- An integer number
n
(type:int
, value:3
) - A string
address
(type:str
, value: Sherlock Holmes' address) - A dictionary
employee
(type:dict
, value: a dictionary which holds three key/value pairs)
Don't worry, I know you're not supposed to know what a dictionary is. We'll see in the next chapter that it's the king of Python data structures.
So, what are n
, address
and employee
? They are names. Names that we can use to retrieve data within our code. They need to be kept somewhere so that whenever we need to retrieve those objects, we can use their names to fetch them. We need some space to hold them, hence: namespaces!
A namespace is therefore a mapping from names to objects. Examples are the set of built-in names (containing functions that are always accessible for free in any Python program), the global names in a module, and the local names in a function. Even the set of attributes of an object can be considered a namespace.
The beauty of namespaces is that they allow you to define and organize your names with clarity, without overlapping or interference. For example, the namespace associated with that book we were looking for in the library can be used to import the book itself, like this:
from library.second_floor.section_x.row_three import book
We start from the library
namespace, and by means of the dot (.
) operator, we walk into that namespace. Within this namespace, we look for second_floor
, and again we walk into it with the .
operator. We then walk into section_x
, and finally within the last namespace, row_three
, we find the name we were looking for: book
.
Walking through a namespace will be clearer when we'll be dealing with real code examples. For now, just keep in mind that namespaces are places where names are associated to objects.
There is another concept, which is closely related to that of a namespace, which I'd like to briefly talk about: the scope.
According to Python's documentation, a scope is a textual region of a Python program, where a namespace is directly accessible. Directly accessible means that when you're looking for an unqualified reference to a name, Python tries to find it in the namespace.
Scopes are determined statically, but actually during runtime they are used dynamically. This means that by inspecting the source code you can tell what the scope of an object is, but this doesn't prevent the software to alter that during runtime. There are four different scopes that Python makes accessible (not necessarily all of them present at the same time, of course):
- The local scope, which is the innermost one and contains the local names.
- The enclosing scope, that is, the scope of any enclosing function. It contains non-local names and also non-global names.
- The global scope contains the global names.
- The built-in scope contains the built-in names. Python comes with a set of functions that you can use in a off-the-shelf fashion, such as
print
,all
,abs
, and so on. They live in the built-in scope.
The rule is the following: when we refer to a name, Python starts looking for it in the current namespace. If the name is not found, Python continues the search to the enclosing scope and this continue until the built-in scope is searched. If a name hasn't been found after searching the built-in scope, then Python raises a NameError
exception, which basically means that the name hasn't been defined (you saw this in the preceding example).
The order in which the namespaces are scanned when looking for a name is therefore: local, enclosing, global, built-in (LEGB).
This is all very theoretical, so let's see an example. In order to show you Local and Enclosing namespaces, I will have to define a few functions. Don't worry if you are not familiar with their syntax for the moment, we'll study functions in Chapter 4, Functions, the Building Blocks of Code. Just remember that in the following code, when you see def
, it means I'm defining a function.
scopes1.py
# Local versus Global # we define a function, called local def local(): m = 7 print(m) m = 5 print(m) # we call, or `execute` the function local local()
In the preceding example, we define the same name m
, both in the global scope and in the local one (the one defined by the function local). When we execute this program with the following command (have you activated your virtualenv?):
$ python scopes1.py
We see two numbers printed on the console: 5
and 7
.
What happens is that the Python interpreter parses the file, top to bottom. First, it finds a couple of comment lines, which are skipped, then it parses the definition of the function local
. When called, this function does two things: it sets up a name to an object representing number 7 and prints it. The Python interpreter keeps going and it finds another name binding. This time the binding happens in the global scope and the value is 5. The next line is a call to the print
function, which is executed (and so we get the first value printed on the console: 5
).
After this, there is a call to the function local
. At this point, Python executes the function, so at this time, the binding m = 7
happens and it's printed.
One very important thing to notice is that the part of the code that belongs to the definition of the function local is indented by four spaces on the right. Python in fact defines scopes by indenting the code. You walk into a scope by indenting and walk out of it by unindenting. Some coders use two spaces, others three, but the suggested number of spaces to use is four. It's a good measure to maximize readability. We'll talk more about all the conventions you should embrace when writing Python code later.
What would happen if we removed that m = 7
line? Remember the LEGB rule. Python would start looking for m
in the local scope (function local
), and, not finding it, it would go to the next enclosing scope. The next one in this case is the global one because there is no enclosing function wrapped around local
. Therefore, we would see two number 5
printed on the console. Let's actually see how the code would look like:
scopes2.py
# Local versus Global
def local():
# m doesn't belong to the scope defined by the local function
# so Python will keep looking into the next enclosing scope.
# m is finally found in the global scope
print(m, 'printing from the local scope')
m = 5
print(m, 'printing from the global scope')
local()
Running scopes2.py
will print this:
(.lpvenv)fab@xps:ch1$ python scopes2.py 5 printing from the global scope 5 printing from the local scope
As expected, Python prints m
the first time, then when the function local
is called, m
isn't found in its scope, so Python looks for it following the LEGB chain until m
is found in the global scope.
Let's see an example with an extra layer, the enclosing scope:
scopes3.py
# Local, Enclosing and Global def enclosing_func(): m = 13 def local(): # m doesn't belong to the scope defined by the local # function so Python will keep looking into the next # enclosing scope. This time m is found in the enclosing # scope print(m, 'printing from the local scope') # calling the function local local() m = 5 print(m, 'printing from the global scope') enclosing_func()
Running scopes3.py
will print on the console:
(.lpvenv)fab@xps:ch1$ python scopes3.py 5 printing from the global scope 13 printing from the local scope
As you can see, the print
instruction from the function local
is referring to m
as before. m
is still not defined within the function itself, so Python starts walking scopes following the LEGB order. This time m
is found in the enclosing scope.
Don't worry if this is still not perfectly clear for now. It will come to you as we go through the examples in the book. The Classes section of the Python tutorial (official documentation) has an interesting paragraph about scopes and namespaces. Make sure you read it at some point if you wish for a deeper understanding of the subject.
Before we finish off this chapter, I would like to talk a bit more about objects. After all, basically everything in Python is an object, so I think they deserve a bit more attention.
When I introduced objects in the A proper introduction section, I said that we use them to represent real-life objects. For example, we sell goods of any kind on the Web nowadays and we need to be able to handle, store, and represent them properly. But objects are actually so much more than that. Most of what you will ever do, in Python, has to do with manipulating objects.
So, without going too much into detail (we'll do that in Chapter 6, Advanced Concepts – OOP, Decorators, and Iterators), I want to give you the in a nutshell kind of explanation about classes and objects.
We've already seen that objects are Python's abstraction for data. In fact, everything in Python is an object. Numbers, strings (data structures that hold text), containers, collections, even functions. You can think of them as if they were boxes with at least three features: an ID (unique), a type, and a value.
But how do they come to life? How do we create them? How to we write our own custom objects? The answer lies in one simple word: classes.
Objects are, in fact, instances of classes. The beauty of Python is that classes are objects themselves, but let's not go down this road. It leads to one of the most advanced concepts of this language: metaclasses. We'll talk very briefly about them in Chapter 6, Advanced Concepts – OOP, Decorators, and Iterators. For now, the best way for you to get the difference between classes and objects, is by means of an example.
Say a friend tells you "I bought a new bike!" You immediately understand what she's talking about. Have you seen the bike? No. Do you know what color it is? Nope. The brand? Nope. Do you know anything about it? Nope. But at the same time, you know everything you need in order to understand what your friend meant when she told you she bought a new bike. You know that a bike has two wheels attached to a frame, a saddle, pedals, handlebars, brakes, and so on. In other words, even if you haven't seen the bike itself, you know the concept of bike. An abstract set of features and characteristics that together form something called bike.
In computer programming, that is called a class. It's that simple. Classes are used to create objects. In fact, objects are said to be instances of classes.
In other words, we all know what a bike is, we know the class. But then I have my own bike, which is an instance of the class bike. And my bike is an object with its own characteristics and methods. You have your own bike. Same class, but different instance. Every bike ever created in the world is an instance of the bike class.
Let's see an example. We will write a class that defines a bike and then we'll create two bikes, one red and one blue. I'll keep the code very simple, but don't fret if you don't understand everything about it; all you need to care about at this moment is to understand the difference between class and object (or instance of a class):
bike.py
# let's define the class Bike class Bike: def __init__(self, colour, frame_material): self.colour = colour self.frame_material = frame_material def brake(self): print("Braking!") # let's create a couple of instances red_bike = Bike('Red', 'Carbon fiber') blue_bike = Bike('Blue', 'Steel') # let's inspect the objects we have, instances of the Bike class. print(red_bike.colour) # prints: Red print(red_bike.frame_material) # prints: Carbon fiber print(blue_bike.colour) # prints: Blue print(blue_bike.frame_material) # prints: Steel # let's brake! red_bike.brake() # prints: Braking!
So many interesting things to notice here. First things first; the definition of a class happens with the class
statement (highlighted in the code). Whatever code comes after the class
statement, and is indented, is called the body of the class. In our case, the last line that belongs to the class definition is the print("Braking!")
one.
After having defined the class we're ready to create instances. You can see that the class body hosts the definition of two methods. A method is basically (and simplistically) a function that belongs to a class.
The first method, __init__
is an initializer. It uses some Python magic to set up the objects with the values we pass when we create it.
Note
Every method that has leading and trailing double underscore, in Python, is called magic method. Magic methods are used by Python for a multitude of different purposes, hence it's never a good idea to name a custom method using two leading and trailing underscores. This naming convention is best left to Python.
The other method we defined, brake
, is just an example of an additional method that we could call if we wanted to brake the bike. It contains just a print
statement, of course, it's an example.
We created two bikes then. One has red color and a carbon fiber frame, and the other one has blue color and steel frame. We pass those values upon creation. After creation, we print out the color property and frame type of the red bike, and the frame type of the blue one just as an example. We also call the brake
method of the red_bike
.
One last thing to notice. You remember I told you that the set of attributes of an object is considered to be a namespace? I hope it's clearer now, what I meant. You see that by getting to the frame_type
property through different namespaces (red_bike
, blue_bike
) we obtain different values. No overlapping, no confusion.
The dot (.
) operator is of course the means we use to walk into a namespace, in the case of objects as well.
Names and namespaces
Say you are looking for a book, so you go to the library and ask someone for the book you want to fetch. They tell you something like "second floor, section X, row three". So you go up the stairs, look for section X, and so on.
It would be very different to enter a library where all the books are piled together in random order in one big room. No floors, no sections, no rows, no order. Fetching a book would be extremely hard.
When we write code we have the same issue: we have to try and organize it so that it will be easy for someone who has no prior knowledge about it to find what they're looking for. When software is structured correctly, it also promotes code reuse. On the other hand, disorganized software is more likely to expose scattered pieces of duplicated logic.
First of all, let's start with the book. We refer to a book by its title and in Python lingo, that would be a name. Python names are the closest abstraction to what other languages call variables. Names basically refer to objects and are introduced by name binding operations. Let's make a quick example (notice that anything that follows a #
is a comment):
>>> n = 3 # integer number >>> address = "221b Baker Street, NW1 6XE, London" # S. Holmes >>> employee = { ... 'age': 45, ... 'role': 'CTO', ... 'SSN': 'AB1234567', ... } >>> # let's print them >>> n 3 >>> address '221b Baker Street, NW1 6XE, London' >>> employee {'role': 'CTO', 'SSN': 'AB1234567', 'age': 45} >>> # what if I try to print a name I didn't define? >>> other_name Traceback (most recent call last): File "<stdin>", line 1, in <module> NameError: name 'other_name' is not defined
We defined three objects in the preceding code (do you remember what are the three features every Python object has?):
- An integer number
n
(type:int
, value:3
) - A string
address
(type:str
, value: Sherlock Holmes' address) - A dictionary
employee
(type:dict
, value: a dictionary which holds three key/value pairs)
Don't worry, I know you're not supposed to know what a dictionary is. We'll see in the next chapter that it's the king of Python data structures.
So, what are n
, address
and employee
? They are names. Names that we can use to retrieve data within our code. They need to be kept somewhere so that whenever we need to retrieve those objects, we can use their names to fetch them. We need some space to hold them, hence: namespaces!
A namespace is therefore a mapping from names to objects. Examples are the set of built-in names (containing functions that are always accessible for free in any Python program), the global names in a module, and the local names in a function. Even the set of attributes of an object can be considered a namespace.
The beauty of namespaces is that they allow you to define and organize your names with clarity, without overlapping or interference. For example, the namespace associated with that book we were looking for in the library can be used to import the book itself, like this:
from library.second_floor.section_x.row_three import book
We start from the library
namespace, and by means of the dot (.
) operator, we walk into that namespace. Within this namespace, we look for second_floor
, and again we walk into it with the .
operator. We then walk into section_x
, and finally within the last namespace, row_three
, we find the name we were looking for: book
.
Walking through a namespace will be clearer when we'll be dealing with real code examples. For now, just keep in mind that namespaces are places where names are associated to objects.
There is another concept, which is closely related to that of a namespace, which I'd like to briefly talk about: the scope.
According to Python's documentation, a scope is a textual region of a Python program, where a namespace is directly accessible. Directly accessible means that when you're looking for an unqualified reference to a name, Python tries to find it in the namespace.
Scopes are determined statically, but actually during runtime they are used dynamically. This means that by inspecting the source code you can tell what the scope of an object is, but this doesn't prevent the software to alter that during runtime. There are four different scopes that Python makes accessible (not necessarily all of them present at the same time, of course):
- The local scope, which is the innermost one and contains the local names.
- The enclosing scope, that is, the scope of any enclosing function. It contains non-local names and also non-global names.
- The global scope contains the global names.
- The built-in scope contains the built-in names. Python comes with a set of functions that you can use in a off-the-shelf fashion, such as
print
,all
,abs
, and so on. They live in the built-in scope.
The rule is the following: when we refer to a name, Python starts looking for it in the current namespace. If the name is not found, Python continues the search to the enclosing scope and this continue until the built-in scope is searched. If a name hasn't been found after searching the built-in scope, then Python raises a NameError
exception, which basically means that the name hasn't been defined (you saw this in the preceding example).
The order in which the namespaces are scanned when looking for a name is therefore: local, enclosing, global, built-in (LEGB).
This is all very theoretical, so let's see an example. In order to show you Local and Enclosing namespaces, I will have to define a few functions. Don't worry if you are not familiar with their syntax for the moment, we'll study functions in Chapter 4, Functions, the Building Blocks of Code. Just remember that in the following code, when you see def
, it means I'm defining a function.
scopes1.py
# Local versus Global # we define a function, called local def local(): m = 7 print(m) m = 5 print(m) # we call, or `execute` the function local local()
In the preceding example, we define the same name m
, both in the global scope and in the local one (the one defined by the function local). When we execute this program with the following command (have you activated your virtualenv?):
$ python scopes1.py
We see two numbers printed on the console: 5
and 7
.
What happens is that the Python interpreter parses the file, top to bottom. First, it finds a couple of comment lines, which are skipped, then it parses the definition of the function local
. When called, this function does two things: it sets up a name to an object representing number 7 and prints it. The Python interpreter keeps going and it finds another name binding. This time the binding happens in the global scope and the value is 5. The next line is a call to the print
function, which is executed (and so we get the first value printed on the console: 5
).
After this, there is a call to the function local
. At this point, Python executes the function, so at this time, the binding m = 7
happens and it's printed.
One very important thing to notice is that the part of the code that belongs to the definition of the function local is indented by four spaces on the right. Python in fact defines scopes by indenting the code. You walk into a scope by indenting and walk out of it by unindenting. Some coders use two spaces, others three, but the suggested number of spaces to use is four. It's a good measure to maximize readability. We'll talk more about all the conventions you should embrace when writing Python code later.
What would happen if we removed that m = 7
line? Remember the LEGB rule. Python would start looking for m
in the local scope (function local
), and, not finding it, it would go to the next enclosing scope. The next one in this case is the global one because there is no enclosing function wrapped around local
. Therefore, we would see two number 5
printed on the console. Let's actually see how the code would look like:
scopes2.py
# Local versus Global
def local():
# m doesn't belong to the scope defined by the local function
# so Python will keep looking into the next enclosing scope.
# m is finally found in the global scope
print(m, 'printing from the local scope')
m = 5
print(m, 'printing from the global scope')
local()
Running scopes2.py
will print this:
(.lpvenv)fab@xps:ch1$ python scopes2.py 5 printing from the global scope 5 printing from the local scope
As expected, Python prints m
the first time, then when the function local
is called, m
isn't found in its scope, so Python looks for it following the LEGB chain until m
is found in the global scope.
Let's see an example with an extra layer, the enclosing scope:
scopes3.py
# Local, Enclosing and Global def enclosing_func(): m = 13 def local(): # m doesn't belong to the scope defined by the local # function so Python will keep looking into the next # enclosing scope. This time m is found in the enclosing # scope print(m, 'printing from the local scope') # calling the function local local() m = 5 print(m, 'printing from the global scope') enclosing_func()
Running scopes3.py
will print on the console:
(.lpvenv)fab@xps:ch1$ python scopes3.py 5 printing from the global scope 13 printing from the local scope
As you can see, the print
instruction from the function local
is referring to m
as before. m
is still not defined within the function itself, so Python starts walking scopes following the LEGB order. This time m
is found in the enclosing scope.
Don't worry if this is still not perfectly clear for now. It will come to you as we go through the examples in the book. The Classes section of the Python tutorial (official documentation) has an interesting paragraph about scopes and namespaces. Make sure you read it at some point if you wish for a deeper understanding of the subject.
Before we finish off this chapter, I would like to talk a bit more about objects. After all, basically everything in Python is an object, so I think they deserve a bit more attention.
When I introduced objects in the A proper introduction section, I said that we use them to represent real-life objects. For example, we sell goods of any kind on the Web nowadays and we need to be able to handle, store, and represent them properly. But objects are actually so much more than that. Most of what you will ever do, in Python, has to do with manipulating objects.
So, without going too much into detail (we'll do that in Chapter 6, Advanced Concepts – OOP, Decorators, and Iterators), I want to give you the in a nutshell kind of explanation about classes and objects.
We've already seen that objects are Python's abstraction for data. In fact, everything in Python is an object. Numbers, strings (data structures that hold text), containers, collections, even functions. You can think of them as if they were boxes with at least three features: an ID (unique), a type, and a value.
But how do they come to life? How do we create them? How to we write our own custom objects? The answer lies in one simple word: classes.
Objects are, in fact, instances of classes. The beauty of Python is that classes are objects themselves, but let's not go down this road. It leads to one of the most advanced concepts of this language: metaclasses. We'll talk very briefly about them in Chapter 6, Advanced Concepts – OOP, Decorators, and Iterators. For now, the best way for you to get the difference between classes and objects, is by means of an example.
Say a friend tells you "I bought a new bike!" You immediately understand what she's talking about. Have you seen the bike? No. Do you know what color it is? Nope. The brand? Nope. Do you know anything about it? Nope. But at the same time, you know everything you need in order to understand what your friend meant when she told you she bought a new bike. You know that a bike has two wheels attached to a frame, a saddle, pedals, handlebars, brakes, and so on. In other words, even if you haven't seen the bike itself, you know the concept of bike. An abstract set of features and characteristics that together form something called bike.
In computer programming, that is called a class. It's that simple. Classes are used to create objects. In fact, objects are said to be instances of classes.
In other words, we all know what a bike is, we know the class. But then I have my own bike, which is an instance of the class bike. And my bike is an object with its own characteristics and methods. You have your own bike. Same class, but different instance. Every bike ever created in the world is an instance of the bike class.
Let's see an example. We will write a class that defines a bike and then we'll create two bikes, one red and one blue. I'll keep the code very simple, but don't fret if you don't understand everything about it; all you need to care about at this moment is to understand the difference between class and object (or instance of a class):
bike.py
# let's define the class Bike class Bike: def __init__(self, colour, frame_material): self.colour = colour self.frame_material = frame_material def brake(self): print("Braking!") # let's create a couple of instances red_bike = Bike('Red', 'Carbon fiber') blue_bike = Bike('Blue', 'Steel') # let's inspect the objects we have, instances of the Bike class. print(red_bike.colour) # prints: Red print(red_bike.frame_material) # prints: Carbon fiber print(blue_bike.colour) # prints: Blue print(blue_bike.frame_material) # prints: Steel # let's brake! red_bike.brake() # prints: Braking!
So many interesting things to notice here. First things first; the definition of a class happens with the class
statement (highlighted in the code). Whatever code comes after the class
statement, and is indented, is called the body of the class. In our case, the last line that belongs to the class definition is the print("Braking!")
one.
After having defined the class we're ready to create instances. You can see that the class body hosts the definition of two methods. A method is basically (and simplistically) a function that belongs to a class.
The first method, __init__
is an initializer. It uses some Python magic to set up the objects with the values we pass when we create it.
Note
Every method that has leading and trailing double underscore, in Python, is called magic method. Magic methods are used by Python for a multitude of different purposes, hence it's never a good idea to name a custom method using two leading and trailing underscores. This naming convention is best left to Python.
The other method we defined, brake
, is just an example of an additional method that we could call if we wanted to brake the bike. It contains just a print
statement, of course, it's an example.
We created two bikes then. One has red color and a carbon fiber frame, and the other one has blue color and steel frame. We pass those values upon creation. After creation, we print out the color property and frame type of the red bike, and the frame type of the blue one just as an example. We also call the brake
method of the red_bike
.
One last thing to notice. You remember I told you that the set of attributes of an object is considered to be a namespace? I hope it's clearer now, what I meant. You see that by getting to the frame_type
property through different namespaces (red_bike
, blue_bike
) we obtain different values. No overlapping, no confusion.
The dot (.
) operator is of course the means we use to walk into a namespace, in the case of objects as well.
Scopes
According to Python's documentation, a scope is a textual region of a Python program, where a namespace is directly accessible. Directly accessible means that when you're looking for an unqualified reference to a name, Python tries to find it in the namespace.
Scopes are determined statically, but actually during runtime they are used dynamically. This means that by inspecting the source code you can tell what the scope of an object is, but this doesn't prevent the software to alter that during runtime. There are four different scopes that Python makes accessible (not necessarily all of them present at the same time, of course):
- The local scope, which is the innermost one and contains the local names.
- The enclosing scope, that is, the scope of any enclosing function. It contains non-local names and also non-global names.
- The global scope contains the global names.
- The built-in scope contains the built-in names. Python comes with a set of functions that you can use in a off-the-shelf fashion, such as
print
,all
,abs
, and so on. They live in the built-in scope.
The rule is the following: when we refer to a name, Python starts looking for it in the current namespace. If the name is not found, Python continues the search to the enclosing scope and this continue until the built-in scope is searched. If a name hasn't been found after searching the built-in scope, then Python raises a NameError
exception, which basically means that the name hasn't been defined (you saw this in the preceding example).
The order in which the namespaces are scanned when looking for a name is therefore: local, enclosing, global, built-in (LEGB).
This is all very theoretical, so let's see an example. In order to show you Local and Enclosing namespaces, I will have to define a few functions. Don't worry if you are not familiar with their syntax for the moment, we'll study functions in Chapter 4, Functions, the Building Blocks of Code. Just remember that in the following code, when you see def
, it means I'm defining a function.
scopes1.py
# Local versus Global # we define a function, called local def local(): m = 7 print(m) m = 5 print(m) # we call, or `execute` the function local local()
In the preceding example, we define the same name m
, both in the global scope and in the local one (the one defined by the function local). When we execute this program with the following command (have you activated your virtualenv?):
$ python scopes1.py
We see two numbers printed on the console: 5
and 7
.
What happens is that the Python interpreter parses the file, top to bottom. First, it finds a couple of comment lines, which are skipped, then it parses the definition of the function local
. When called, this function does two things: it sets up a name to an object representing number 7 and prints it. The Python interpreter keeps going and it finds another name binding. This time the binding happens in the global scope and the value is 5. The next line is a call to the print
function, which is executed (and so we get the first value printed on the console: 5
).
After this, there is a call to the function local
. At this point, Python executes the function, so at this time, the binding m = 7
happens and it's printed.
One very important thing to notice is that the part of the code that belongs to the definition of the function local is indented by four spaces on the right. Python in fact defines scopes by indenting the code. You walk into a scope by indenting and walk out of it by unindenting. Some coders use two spaces, others three, but the suggested number of spaces to use is four. It's a good measure to maximize readability. We'll talk more about all the conventions you should embrace when writing Python code later.
What would happen if we removed that m = 7
line? Remember the LEGB rule. Python would start looking for m
in the local scope (function local
), and, not finding it, it would go to the next enclosing scope. The next one in this case is the global one because there is no enclosing function wrapped around local
. Therefore, we would see two number 5
printed on the console. Let's actually see how the code would look like:
scopes2.py
# Local versus Global
def local():
# m doesn't belong to the scope defined by the local function
# so Python will keep looking into the next enclosing scope.
# m is finally found in the global scope
print(m, 'printing from the local scope')
m = 5
print(m, 'printing from the global scope')
local()
Running scopes2.py
will print this:
(.lpvenv)fab@xps:ch1$ python scopes2.py 5 printing from the global scope 5 printing from the local scope
As expected, Python prints m
the first time, then when the function local
is called, m
isn't found in its scope, so Python looks for it following the LEGB chain until m
is found in the global scope.
Let's see an example with an extra layer, the enclosing scope:
scopes3.py
# Local, Enclosing and Global def enclosing_func(): m = 13 def local(): # m doesn't belong to the scope defined by the local # function so Python will keep looking into the next # enclosing scope. This time m is found in the enclosing # scope print(m, 'printing from the local scope') # calling the function local local() m = 5 print(m, 'printing from the global scope') enclosing_func()
Running scopes3.py
will print on the console:
(.lpvenv)fab@xps:ch1$ python scopes3.py 5 printing from the global scope 13 printing from the local scope
As you can see, the print
instruction from the function local
is referring to m
as before. m
is still not defined within the function itself, so Python starts walking scopes following the LEGB order. This time m
is found in the enclosing scope.
Don't worry if this is still not perfectly clear for now. It will come to you as we go through the examples in the book. The Classes section of the Python tutorial (official documentation) has an interesting paragraph about scopes and namespaces. Make sure you read it at some point if you wish for a deeper understanding of the subject.
Before we finish off this chapter, I would like to talk a bit more about objects. After all, basically everything in Python is an object, so I think they deserve a bit more attention.
When I introduced objects in the A proper introduction section, I said that we use them to represent real-life objects. For example, we sell goods of any kind on the Web nowadays and we need to be able to handle, store, and represent them properly. But objects are actually so much more than that. Most of what you will ever do, in Python, has to do with manipulating objects.
So, without going too much into detail (we'll do that in Chapter 6, Advanced Concepts – OOP, Decorators, and Iterators), I want to give you the in a nutshell kind of explanation about classes and objects.
We've already seen that objects are Python's abstraction for data. In fact, everything in Python is an object. Numbers, strings (data structures that hold text), containers, collections, even functions. You can think of them as if they were boxes with at least three features: an ID (unique), a type, and a value.
But how do they come to life? How do we create them? How to we write our own custom objects? The answer lies in one simple word: classes.
Objects are, in fact, instances of classes. The beauty of Python is that classes are objects themselves, but let's not go down this road. It leads to one of the most advanced concepts of this language: metaclasses. We'll talk very briefly about them in Chapter 6, Advanced Concepts – OOP, Decorators, and Iterators. For now, the best way for you to get the difference between classes and objects, is by means of an example.
Say a friend tells you "I bought a new bike!" You immediately understand what she's talking about. Have you seen the bike? No. Do you know what color it is? Nope. The brand? Nope. Do you know anything about it? Nope. But at the same time, you know everything you need in order to understand what your friend meant when she told you she bought a new bike. You know that a bike has two wheels attached to a frame, a saddle, pedals, handlebars, brakes, and so on. In other words, even if you haven't seen the bike itself, you know the concept of bike. An abstract set of features and characteristics that together form something called bike.
In computer programming, that is called a class. It's that simple. Classes are used to create objects. In fact, objects are said to be instances of classes.
In other words, we all know what a bike is, we know the class. But then I have my own bike, which is an instance of the class bike. And my bike is an object with its own characteristics and methods. You have your own bike. Same class, but different instance. Every bike ever created in the world is an instance of the bike class.
Let's see an example. We will write a class that defines a bike and then we'll create two bikes, one red and one blue. I'll keep the code very simple, but don't fret if you don't understand everything about it; all you need to care about at this moment is to understand the difference between class and object (or instance of a class):
bike.py
# let's define the class Bike class Bike: def __init__(self, colour, frame_material): self.colour = colour self.frame_material = frame_material def brake(self): print("Braking!") # let's create a couple of instances red_bike = Bike('Red', 'Carbon fiber') blue_bike = Bike('Blue', 'Steel') # let's inspect the objects we have, instances of the Bike class. print(red_bike.colour) # prints: Red print(red_bike.frame_material) # prints: Carbon fiber print(blue_bike.colour) # prints: Blue print(blue_bike.frame_material) # prints: Steel # let's brake! red_bike.brake() # prints: Braking!
So many interesting things to notice here. First things first; the definition of a class happens with the class
statement (highlighted in the code). Whatever code comes after the class
statement, and is indented, is called the body of the class. In our case, the last line that belongs to the class definition is the print("Braking!")
one.
After having defined the class we're ready to create instances. You can see that the class body hosts the definition of two methods. A method is basically (and simplistically) a function that belongs to a class.
The first method, __init__
is an initializer. It uses some Python magic to set up the objects with the values we pass when we create it.
Note
Every method that has leading and trailing double underscore, in Python, is called magic method. Magic methods are used by Python for a multitude of different purposes, hence it's never a good idea to name a custom method using two leading and trailing underscores. This naming convention is best left to Python.
The other method we defined, brake
, is just an example of an additional method that we could call if we wanted to brake the bike. It contains just a print
statement, of course, it's an example.
We created two bikes then. One has red color and a carbon fiber frame, and the other one has blue color and steel frame. We pass those values upon creation. After creation, we print out the color property and frame type of the red bike, and the frame type of the blue one just as an example. We also call the brake
method of the red_bike
.
One last thing to notice. You remember I told you that the set of attributes of an object is considered to be a namespace? I hope it's clearer now, what I meant. You see that by getting to the frame_type
property through different namespaces (red_bike
, blue_bike
) we obtain different values. No overlapping, no confusion.
The dot (.
) operator is of course the means we use to walk into a namespace, in the case of objects as well.
Object and classes
When I introduced objects in the A proper introduction section, I said that we use them to represent real-life objects. For example, we sell goods of any kind on the Web nowadays and we need to be able to handle, store, and represent them properly. But objects are actually so much more than that. Most of what you will ever do, in Python, has to do with manipulating objects.
So, without going too much into detail (we'll do that in Chapter 6, Advanced Concepts – OOP, Decorators, and Iterators), I want to give you the in a nutshell kind of explanation about classes and objects.
We've already seen that objects are Python's abstraction for data. In fact, everything in Python is an object. Numbers, strings (data structures that hold text), containers, collections, even functions. You can think of them as if they were boxes with at least three features: an ID (unique), a type, and a value.
But how do they come to life? How do we create them? How to we write our own custom objects? The answer lies in one simple word: classes.
Objects are, in fact, instances of classes. The beauty of Python is that classes are objects themselves, but let's not go down this road. It leads to one of the most advanced concepts of this language: metaclasses. We'll talk very briefly about them in Chapter 6, Advanced Concepts – OOP, Decorators, and Iterators. For now, the best way for you to get the difference between classes and objects, is by means of an example.
Say a friend tells you "I bought a new bike!" You immediately understand what she's talking about. Have you seen the bike? No. Do you know what color it is? Nope. The brand? Nope. Do you know anything about it? Nope. But at the same time, you know everything you need in order to understand what your friend meant when she told you she bought a new bike. You know that a bike has two wheels attached to a frame, a saddle, pedals, handlebars, brakes, and so on. In other words, even if you haven't seen the bike itself, you know the concept of bike. An abstract set of features and characteristics that together form something called bike.
In computer programming, that is called a class. It's that simple. Classes are used to create objects. In fact, objects are said to be instances of classes.
In other words, we all know what a bike is, we know the class. But then I have my own bike, which is an instance of the class bike. And my bike is an object with its own characteristics and methods. You have your own bike. Same class, but different instance. Every bike ever created in the world is an instance of the bike class.
Let's see an example. We will write a class that defines a bike and then we'll create two bikes, one red and one blue. I'll keep the code very simple, but don't fret if you don't understand everything about it; all you need to care about at this moment is to understand the difference between class and object (or instance of a class):
bike.py
# let's define the class Bike class Bike: def __init__(self, colour, frame_material): self.colour = colour self.frame_material = frame_material def brake(self): print("Braking!") # let's create a couple of instances red_bike = Bike('Red', 'Carbon fiber') blue_bike = Bike('Blue', 'Steel') # let's inspect the objects we have, instances of the Bike class. print(red_bike.colour) # prints: Red print(red_bike.frame_material) # prints: Carbon fiber print(blue_bike.colour) # prints: Blue print(blue_bike.frame_material) # prints: Steel # let's brake! red_bike.brake() # prints: Braking!
So many interesting things to notice here. First things first; the definition of a class happens with the class
statement (highlighted in the code). Whatever code comes after the class
statement, and is indented, is called the body of the class. In our case, the last line that belongs to the class definition is the print("Braking!")
one.
After having defined the class we're ready to create instances. You can see that the class body hosts the definition of two methods. A method is basically (and simplistically) a function that belongs to a class.
The first method, __init__
is an initializer. It uses some Python magic to set up the objects with the values we pass when we create it.
Note
Every method that has leading and trailing double underscore, in Python, is called magic method. Magic methods are used by Python for a multitude of different purposes, hence it's never a good idea to name a custom method using two leading and trailing underscores. This naming convention is best left to Python.
The other method we defined, brake
, is just an example of an additional method that we could call if we wanted to brake the bike. It contains just a print
statement, of course, it's an example.
We created two bikes then. One has red color and a carbon fiber frame, and the other one has blue color and steel frame. We pass those values upon creation. After creation, we print out the color property and frame type of the red bike, and the frame type of the blue one just as an example. We also call the brake
method of the red_bike
.
One last thing to notice. You remember I told you that the set of attributes of an object is considered to be a namespace? I hope it's clearer now, what I meant. You see that by getting to the frame_type
property through different namespaces (red_bike
, blue_bike
) we obtain different values. No overlapping, no confusion.
The dot (.
) operator is of course the means we use to walk into a namespace, in the case of objects as well.
Writing good code is not as easy as it seems. As I already said before, good code exposes a long list of qualities that is quite hard to put together. Writing good code is, to some extent, an art. Regardless of where on the path you will be happy to settle, there is something that you can embrace which will make your code instantly better: PEP8.
According to Wikipedia:
"Python's development is conducted largely through the Python Enhancement Proposal (PEP) process. The PEP process is the primary mechanism for proposing major new features, for collecting community input on an issue, and for documenting the design decisions that have gone into Python."
Among all the PEPs, probably the most famous one is PEP8. It lays out a simple but effective set of guidelines to define Python aesthetic so that we write beautiful Python code. If you take one suggestion out of this chapter, please let it be this: use it. Embrace it. You will thank me later.
Coding today is no longer a check-in/check-out business. Rather, it's more of a social effort. Several developers collaborate to a piece of code through tools like git and mercurial, and the result is code that is fathered by many different hands.
These days, more than ever, we need to have a consistent way of writing code, so that readability is maximized. When all developers of a company abide with PEP8, it's not uncommon for any of them landing on a piece of code to think they wrote it themselves. It actually happens to me all the time (I always forget the code I write).
This has a tremendous advantage: when you read code that you could have written yourself, you read it easily. Without a convention, every coder would structure the code the way they like most, or simply the way they were taught or are used to, and this would mean having to interpret every line according to someone else's style. It would mean having to lose much more time just trying to understand it. Thanks to PEP8, we can avoid this. I'm such a fan of it that I won't sign off a code review if the code doesn't respect it. So please take the time to study it, it's very important.
In the examples of this book, I will try to respect it as much as I can. Unfortunately, I don't have the luxury of 79 characters (which is the maximum line length suggested by PEP*), and I will have to cut down on blank lines and other things, but I promise you I'll try to layout my code so that it's as readable as possible.
Python has been adopted widely in all coding industries. It's used by many different companies for many different purposes, and it's also used in education (it's an excellent language for that purpose, because of its many qualities and the fact that it's easy to learn).
One of the reasons Python is so popular today is that the community around it is vast, vibrant, and full of brilliant people. Many events are organized all over the world, mostly either around Python or its main web framework, Django.
Python is open, and very often so are the minds of those who embrace it. Check out the community page on the Python website for more information and get involved!
There is another aspect to Python which revolves around the notion of being Pythonic. It has to do with the fact that Python allows you to use some idioms that aren't found elsewhere, at least not in the same form or easiness of use (I feel quite claustrophobic when I have to code in a language which is not Python now).
Anyway, over the years, this concept of being Pythonic has emerged and, the way I understand it, is something along the lines of doing things the way they are supposed to be done in Python.
To help you understand a little bit more about Python's culture and about being Pythonic, I will show you the Zen of Python. A lovely Easter egg that is very popular. Open up a Python console and type import this
. What follows is the result of this line:
>>> import this The Zen of Python, by Tim Peters Beautiful is better than ugly. Explicit is better than implicit. Simple is better than complex. Complex is better than complicated. Flat is better than nested. Sparse is better than dense. Readability counts. Special cases aren't special enough to break the rules. Although practicality beats purity. Errors should never pass silently. Unless explicitly silenced. In the face of ambiguity, refuse the temptation to guess. There should be one-- and preferably only one --obvious way to do it. Although that way may not be obvious at first unless you're Dutch. Now is better than never. Although never is often better than *right* now. If the implementation is hard to explain, it's a bad idea. If the implementation is easy to explain, it may be a good idea. Namespaces are one honking great idea -- let's do more of those!
There are two levels of reading here. One is to consider it as a set of guidelines that have been put down in a fun way. The other one is to keep it in mind, and maybe read it once in a while, trying to understand how it refers to something deeper. Some Python characteristics that you will have to understand deeply in order to write Python the way it's supposed to be written. Start with the fun level, and then dig deeper. Always dig deeper.
Just a few words about Integrated Development Environments (IDEs). To follow the examples in this book you don't need one, any text editor will do fine. If you want to have more advanced features such as syntax coloring and auto completion, you will have to fetch yourself an IDE. You can find a comprehensive list of open source IDEs (just Google "python ides") on the Python website. I personally use Sublime Text editor. It's free to try out and it costs just a few dollars. I have tried many IDEs in my life, but this is the one that makes me most productive.
Two extremely important pieces of advice:
- Whatever IDE you will chose to use, try to learn it well so that you can exploit its strengths, but don't depend on it. Exercise yourself to work with VIM (or any other text editor) once in a while, learn to be able to do some work on any platform, with any set of tools.
- Whatever text editor/IDE you will use, when it comes to writing Python, indentation is four spaces. Don't use tabs, don't mix them with spaces. Use four spaces, not two, not three, not five. Just use four. The whole world works like that, and you don't want to become an outcast because you were fond of the three-space layout.
In this chapter, we started to explore the world of programming and that of Python. We've barely scratched the surface, just a little, touching concepts that will be discussed later on in the book in greater detail.
We talked about Python's main features, who is using it and for what, and what are the different ways in which we can write a Python program.
In the last part of the chapter, we flew over the fundamental notions of namespace, scope, class, and object. We also saw how Python code can be organized using modules and packages.
On a practical level, we learned how to install Python on our system, how to make sure we have the tools we need, pip and virtualenv, and we also created and activated our first virtual environment. This will allow us to work in a self-contained environment without the risk of compromising the Python system installation.
Now you're ready to start this journey with me. All you need is enthusiasm, an activated virtual environment, this book, your fingers, and some coffee.
Try to follow the examples, I'll keep them simple and short. If you put them under your fingertips, you will retain them much better than if you just read them.
In the next chapter, we will explore Python's rich set of built-in data types. There's much to cover and much to learn!
"Data! Data! Data!" he cried impatiently. "I can't make bricks without clay." | ||
--Sherlock Holmes - The Adventure of the Copper Beeches |
Everything you do with a computer is managing data. Data comes in many different shapes and flavors. It's the music you listen, the movie you stream, the PDFs you open. Even the chapter you're reading at this very moment is just a file, which is data.
Data can be simple, an integer number to represent an age, or complex, like an order placed on a website. It can be about a single object or about a collection of them.
Data can even be about data, that is, metadata. Data that describes the design of other data structures or data that describes application data or its context.
In Python, objects are abstraction for data, and Python has an amazing variety of data structures that you can use to represent data, or combine them to create your own custom data. Before we delve into the specifics, I want you to be very clear about objects in Python, so let's talk a little bit more about them.
As we already said, everything in Python is an object. But what really happens when you type an instruction like age = 42
in a Python module?
Tip
If you go to http://pythontutor.com/, you can type that instruction into a text box and get its visual representation. Keep this website in mind, it's very useful to consolidate your understanding of what goes on behind the scenes.
So, what happens is that an object is created. It gets an id
, the type
is set to int
(integer number), and the value
to 42
. A name age
is placed in the global namespace, pointing to that object. Therefore, whenever we are in the global namespace, after the execution of that line, we can retrieve that object by simply accessing it through its name: age
.
If you were to move house, you would put all the knives, forks, and spoons in a box and label it cutlery. Can you see it's exactly the same concept? Here's a screenshot of how it may look like (you may have to tweak the settings to get to the same view):

So, for the rest of this chapter, whenever you read something such as name = some_value
, think of a name placed in the namespace that is tied to the scope in which the instruction was written, with a nice arrow pointing to an object that has an id
, a type
, and a value
. There is a little bit more to say about this mechanism, but it's much easier to talk about it over an example, so we'll get back to this later.
A first fundamental distinction that Python makes on data is about whether or not the value of an object changes. If the value can change, the object is called mutable, while if the value cannot change, the object is called immutable.
It is very important that you understand the distinction between mutable and immutable because it affects the code you write, so here's a question:
>>> age = 42 >>> age 42 >>> age = 43 #A >>> age 43
In the preceding code, on the line #A
, have I changed the value of age? Well, no. But now it's 43 (I hear you say...). Yes, it's 43, but 42 was an integer number, of the type int
, which is immutable. So, what happened is really that on the first line, age
is a name that is set to point to an int
object, whose value is 42. When we type age = 43
, what happens is that another object is created, of the type int
and value 43 (also, the id
will be different), and the name age
is set to point to it. So, we didn't change that 42 to 43. We actually just pointed age
to a different location: the new int
object whose value is 43. Let's see the same code also printing the IDs:
>>> age = 42 >>> id(age) 10456352 >>> age = 43 >>> id(age) 10456384
Notice that we print the IDs by calling the built-in id
function. As you can see, they are different, as expected. Bear in mind that age
points to one object at a time: 42 first, then 43. Never together.
Now, let's see the same example using a mutable object. For this example, let's just use a Person
object, that has a property age
:
>>> fab = Person(age=39) >>> fab.age 39 >>> id(fab) 139632387887456 >>> fab.age = 29 # I wish! >>> id(fab) 139632387887456 # still the same id
In this case, I set up an object fab
whose type
is Person
(a custom class). On creation, the object is given the age
of 39. I'm printing it, along with the object id
, right afterwards. Notice that, even after I change age
to be 29, the ID of fab
stays the same (while the ID of age
has changed, of course). Custom objects in Python are mutable (unless you code them not to be). Keep this concept in mind, it's very important. I'll remind you about it through the rest of the chapter.
Let's start by exploring Python's built-in data types for numbers. Python was designed by a man with a master's degree in mathematics and computer science, so it's only logical that it has amazing support for numbers.
Numbers are immutable objects.
Python integers have unlimited range, subject only to the available virtual memory. This means that it doesn't really matter how big a number you want to store: as long as it can fit in your computer's memory, Python will take care of it. Integer numbers can be positive, negative, and 0 (zero). They support all the basic mathematical operations, as shown in the following example:
>>> a = 12 >>> b = 3 >>> a + b # addition 15 >>> b - a # subtraction -9 >>> a // b # integer division 4 >>> a / b # true division 4.0 >>> a * b # multiplication 36 >>> b ** a # power operator 531441 >>> 2 ** 1024 # a very big number, Python handles it gracefully 17976931348623159077293051907890247336179769789423065727343008115 77326758055009631327084773224075360211201138798713933576587897688 14416622492847430639474124377767893424865485276302219601246094119 45308295208500576883815068234246288147391311054082723716335051068 4586298239947245938479716304835356329624224137216
The preceding code should be easy to understand. Just notice one important thing: Python has two division operators, one performs the so-called true division (/
), which returns the quotient of the operands, and the other one, the so-called integer division (//
), which returns the floored quotient of the operands. See how that is different for positive and negative numbers:
>>> 7 / 4 # true division 1.75 >>> 7 // 4 # integer division, flooring returns 1 1 >>> -7 / 4 # true division again, result is opposite of previous -1.75 >>> -7 // 4 # integer div., result not the opposite of previous -2
This is an interesting example. If you were expecting a -1
on the last line, don't feel bad, it's just the way Python works. The result of an integer division in Python is always rounded towards minus infinity. If instead of flooring you want to truncate a number to an integer, you can use the built-in int
function, like shown in the following example:
>>> int(1.75) 1 >>> int(-1.75) -1
Notice that truncation is done towards 0.
There is also an operator to calculate the remainder of a division. It's called modulo operator, and it's represented by a percent (%
):
>>> 10 % 3 # remainder of the division 10 // 3 1 >>> 10 % 4 # remainder of the division 10 // 4 2
Boolean algebra is that subset of algebra in which the values of the variables are the truth values: true and false. In Python, True
and False
are two keywords that are used to represent truth values. Booleans are a subclass of integers, and behave respectively like 1 and 0. The equivalent of the int
class for Booleans is the bool
class, which returns either True
or False
. Every built-in Python object has a value in the Boolean context, which means they basically evaluate to either True
or False
when fed to the bool
function. We'll see all about this in Chapter 3, Iterating and Making Decisions.
Boolean values can be combined in Boolean expressions using the logical operators and
, or
, and not
. Again, we'll see them in full in the next chapter, so for now let's just see a simple example:
>>> int(True) # True behaves like 1 1 >>> int(False) # False behaves like 0 0 >>> bool(1) # 1 evaluates to True in a boolean context True >>> bool(-42) # and so does every non-zero number True >>> bool(0) # 0 evaluates to False False >>> # quick peak at the operators (and, or, not) >>> not True False >>> not False True >>> True and True True >>> False or True True
You can see that True
and False
are subclasses of integers when you try to add them. Python upcasts them to integers and performs addition:
>>> 1 + True 2 >>> False + 42 42 >>> 7 - True 6
Note
Upcasting is a type conversion operation that goes from a subclass to its parent. In the example presented here, True
and False
, which belong to a class derived from the integer class, are converted back to integers when needed. This topic is about inheritance and will be explained in detail in Chapter 6, Advanced Concepts – OOP, Decorators, and Iterators.
Real numbers, or floating point numbers, are represented in Python according to the IEEE 754 double-precision binary floating-point format, which is stored in 64 bits of information divided into three sections: sign, exponent, and mantissa.
Note
Quench your thirst for knowledge about this format on Wikipedia: http://en.wikipedia.org/wiki/Double-precision_floating-point_format
Usually programming languages give coders two different formats: single and double precision. The former taking up 32 bits of memory, and the latter 64. Python supports only the double format. Let's see a simple example:
>>> pi = 3.1415926536 # how many digits of PI can you remember? >>> radius = 4.5 >>> area = pi * (radius ** 2) >>> area 63.61725123519331
The sys.float_info
struct sequence holds information about how floating point numbers will behave on your system. This is what I see on my box:
>>> import sys >>> sys.float_info sys.float_info(max=1.7976931348623157e+308, max_exp=1024, max_10_exp=308, min=2.2250738585072014e-308, min_exp=-1021, min_10_exp=-307, dig=15, mant_dig=53, epsilon=2.220446049250313e-16, radix=2, rounds=1)
Let's make a few considerations here: we have 64 bits to represent float numbers. This means we can represent at most 2 ** 64 == 18,446,744,073,709,551,616
numbers with that amount of bits. Take a look at the max
and epsilon
value for the float numbers, and you'll realize it's impossible to represent them all. There is just not enough space so they are approximated to the closest representable number. You probably think that only extremely big or extremely small numbers suffer from this issue. Well, think again:
>>> 3 * 0.1 – 0.3 # this should be 0!!! 5.551115123125783e-17
What does this tell you? It tells you that double precision numbers suffer from approximation issues even when it comes to simple numbers like 0.1 or 0.3. Why is this important? It can be a big problem if you're handling prices, or financial calculations, or any kind of data that needs not to be approximated. Don't worry, Python gives you the Decimal type, which doesn't suffer from these issues, we'll see them in a bit.
Python gives you complex numbers support out of the box. If you don't know what complex numbers are, you can look them up on the Web. They are numbers that can be expressed in the form a + ib where a and b are real numbers, and i (or j if you're an engineer) is the imaginary unit, that is, the square root of -1. a and b are called respectively the real and imaginary part of the number.
It's actually unlikely you'll be using them, unless you're coding something scientific. Let's see a small example:
>>> c = 3.14 + 2.73j >>> c.real # real part 3.14 >>> c.imag # imaginary part 2.73 >>> c.conjugate() # conjugate of A + Bj is A - Bj (3.14-2.73j) >>> c * 2 # multiplication is allowed (6.28+5.46j) >>> c ** 2 # power operation as well (2.4067000000000007+17.1444j) >>> d = 1 + 1j # addition and subtraction as well >>> c - d (2.14+1.73j)
Let's finish the tour of the number department with a look at fractions and decimals. Fractions hold a rational numerator and denominator in their lowest forms. Let's see a quick example:
>>> from fractions import Fraction >>> Fraction(10, 6) # mad hatter? Fraction(5, 3) # notice it's been reduced to lowest terms >>> Fraction(1, 3) + Fraction(2, 3) # 1/3 + 2/3 = 3/3 = 1/1 Fraction(1, 1) >>> f = Fraction(10, 6) >>> f.numerator 5 >>> f.denominator 3
Although they can be very useful at times, it's not that common to spot them in commercial software. Much easier instead, is to see decimal numbers being used in all those contexts where precision is everything, for example, scientific and financial calculations.
Note
It's important to remember that arbitrary precision decimal numbers come at a price in performance, of course. The amount of data to be stored for each number is far greater than it is for fractions or floats as well as the way they are handled, which requires the Python interpreter much more work behind the scenes. Another interesting thing to know is that you can get and set the precision by accessing decimal.getcontext().prec
.
Let's see a quick example with Decimal
numbers:
>>> from decimal import Decimal as D # rename for brevity >>> D(3.14) # pi, from float, so approximation issues Decimal('3.140000000000000124344978758017532527446746826171875') >>> D('3.14') # pi, from a string, so no approximation issues Decimal('3.14') >>> D(0.1) * D(3) - D(0.3) # from float, we still have the issue Decimal('2.775557561565156540423631668E-17') >>> D('0.1') * D(3) - D('0.3') # from string, all perfect Decimal('0.0')
Notice that when we construct a Decimal
number from a float
, it takes on all the approximation issues the float
may come from. On the other hand, when the Decimal
has no approximation issues, for example, when we feed an int
or a string
representation to the constructor, then the calculation has no quirky behavior. When it comes to money, use decimals.
This concludes our introduction to built-in numeric types, let's now see sequences.
Integers
Python integers have unlimited range, subject only to the available virtual memory. This means that it doesn't really matter how big a number you want to store: as long as it can fit in your computer's memory, Python will take care of it. Integer numbers can be positive, negative, and 0 (zero). They support all the basic mathematical operations, as shown in the following example:
>>> a = 12 >>> b = 3 >>> a + b # addition 15 >>> b - a # subtraction -9 >>> a // b # integer division 4 >>> a / b # true division 4.0 >>> a * b # multiplication 36 >>> b ** a # power operator 531441 >>> 2 ** 1024 # a very big number, Python handles it gracefully 17976931348623159077293051907890247336179769789423065727343008115 77326758055009631327084773224075360211201138798713933576587897688 14416622492847430639474124377767893424865485276302219601246094119 45308295208500576883815068234246288147391311054082723716335051068 4586298239947245938479716304835356329624224137216
The preceding code should be easy to understand. Just notice one important thing: Python has two division operators, one performs the so-called true division (/
), which returns the quotient of the operands, and the other one, the so-called integer division (//
), which returns the floored quotient of the operands. See how that is different for positive and negative numbers:
>>> 7 / 4 # true division 1.75 >>> 7 // 4 # integer division, flooring returns 1 1 >>> -7 / 4 # true division again, result is opposite of previous -1.75 >>> -7 // 4 # integer div., result not the opposite of previous -2
This is an interesting example. If you were expecting a -1
on the last line, don't feel bad, it's just the way Python works. The result of an integer division in Python is always rounded towards minus infinity. If instead of flooring you want to truncate a number to an integer, you can use the built-in int
function, like shown in the following example:
>>> int(1.75) 1 >>> int(-1.75) -1
Notice that truncation is done towards 0.
There is also an operator to calculate the remainder of a division. It's called modulo operator, and it's represented by a percent (%
):
>>> 10 % 3 # remainder of the division 10 // 3 1 >>> 10 % 4 # remainder of the division 10 // 4 2
Boolean algebra is that subset of algebra in which the values of the variables are the truth values: true and false. In Python, True
and False
are two keywords that are used to represent truth values. Booleans are a subclass of integers, and behave respectively like 1 and 0. The equivalent of the int
class for Booleans is the bool
class, which returns either True
or False
. Every built-in Python object has a value in the Boolean context, which means they basically evaluate to either True
or False
when fed to the bool
function. We'll see all about this in Chapter 3, Iterating and Making Decisions.
Boolean values can be combined in Boolean expressions using the logical operators and
, or
, and not
. Again, we'll see them in full in the next chapter, so for now let's just see a simple example:
>>> int(True) # True behaves like 1 1 >>> int(False) # False behaves like 0 0 >>> bool(1) # 1 evaluates to True in a boolean context True >>> bool(-42) # and so does every non-zero number True >>> bool(0) # 0 evaluates to False False >>> # quick peak at the operators (and, or, not) >>> not True False >>> not False True >>> True and True True >>> False or True True
You can see that True
and False
are subclasses of integers when you try to add them. Python upcasts them to integers and performs addition:
>>> 1 + True 2 >>> False + 42 42 >>> 7 - True 6
Note
Upcasting is a type conversion operation that goes from a subclass to its parent. In the example presented here, True
and False
, which belong to a class derived from the integer class, are converted back to integers when needed. This topic is about inheritance and will be explained in detail in Chapter 6, Advanced Concepts – OOP, Decorators, and Iterators.
Real numbers, or floating point numbers, are represented in Python according to the IEEE 754 double-precision binary floating-point format, which is stored in 64 bits of information divided into three sections: sign, exponent, and mantissa.
Note
Quench your thirst for knowledge about this format on Wikipedia: http://en.wikipedia.org/wiki/Double-precision_floating-point_format
Usually programming languages give coders two different formats: single and double precision. The former taking up 32 bits of memory, and the latter 64. Python supports only the double format. Let's see a simple example:
>>> pi = 3.1415926536 # how many digits of PI can you remember? >>> radius = 4.5 >>> area = pi * (radius ** 2) >>> area 63.61725123519331
The sys.float_info
struct sequence holds information about how floating point numbers will behave on your system. This is what I see on my box:
>>> import sys >>> sys.float_info sys.float_info(max=1.7976931348623157e+308, max_exp=1024, max_10_exp=308, min=2.2250738585072014e-308, min_exp=-1021, min_10_exp=-307, dig=15, mant_dig=53, epsilon=2.220446049250313e-16, radix=2, rounds=1)
Let's make a few considerations here: we have 64 bits to represent float numbers. This means we can represent at most 2 ** 64 == 18,446,744,073,709,551,616
numbers with that amount of bits. Take a look at the max
and epsilon
value for the float numbers, and you'll realize it's impossible to represent them all. There is just not enough space so they are approximated to the closest representable number. You probably think that only extremely big or extremely small numbers suffer from this issue. Well, think again:
>>> 3 * 0.1 – 0.3 # this should be 0!!! 5.551115123125783e-17
What does this tell you? It tells you that double precision numbers suffer from approximation issues even when it comes to simple numbers like 0.1 or 0.3. Why is this important? It can be a big problem if you're handling prices, or financial calculations, or any kind of data that needs not to be approximated. Don't worry, Python gives you the Decimal type, which doesn't suffer from these issues, we'll see them in a bit.
Python gives you complex numbers support out of the box. If you don't know what complex numbers are, you can look them up on the Web. They are numbers that can be expressed in the form a + ib where a and b are real numbers, and i (or j if you're an engineer) is the imaginary unit, that is, the square root of -1. a and b are called respectively the real and imaginary part of the number.
It's actually unlikely you'll be using them, unless you're coding something scientific. Let's see a small example:
>>> c = 3.14 + 2.73j >>> c.real # real part 3.14 >>> c.imag # imaginary part 2.73 >>> c.conjugate() # conjugate of A + Bj is A - Bj (3.14-2.73j) >>> c * 2 # multiplication is allowed (6.28+5.46j) >>> c ** 2 # power operation as well (2.4067000000000007+17.1444j) >>> d = 1 + 1j # addition and subtraction as well >>> c - d (2.14+1.73j)
Let's finish the tour of the number department with a look at fractions and decimals. Fractions hold a rational numerator and denominator in their lowest forms. Let's see a quick example:
>>> from fractions import Fraction >>> Fraction(10, 6) # mad hatter? Fraction(5, 3) # notice it's been reduced to lowest terms >>> Fraction(1, 3) + Fraction(2, 3) # 1/3 + 2/3 = 3/3 = 1/1 Fraction(1, 1) >>> f = Fraction(10, 6) >>> f.numerator 5 >>> f.denominator 3
Although they can be very useful at times, it's not that common to spot them in commercial software. Much easier instead, is to see decimal numbers being used in all those contexts where precision is everything, for example, scientific and financial calculations.
Note
It's important to remember that arbitrary precision decimal numbers come at a price in performance, of course. The amount of data to be stored for each number is far greater than it is for fractions or floats as well as the way they are handled, which requires the Python interpreter much more work behind the scenes. Another interesting thing to know is that you can get and set the precision by accessing decimal.getcontext().prec
.
Let's see a quick example with Decimal
numbers:
>>> from decimal import Decimal as D # rename for brevity >>> D(3.14) # pi, from float, so approximation issues Decimal('3.140000000000000124344978758017532527446746826171875') >>> D('3.14') # pi, from a string, so no approximation issues Decimal('3.14') >>> D(0.1) * D(3) - D(0.3) # from float, we still have the issue Decimal('2.775557561565156540423631668E-17') >>> D('0.1') * D(3) - D('0.3') # from string, all perfect Decimal('0.0')
Notice that when we construct a Decimal
number from a float
, it takes on all the approximation issues the float
may come from. On the other hand, when the Decimal
has no approximation issues, for example, when we feed an int
or a string
representation to the constructor, then the calculation has no quirky behavior. When it comes to money, use decimals.
This concludes our introduction to built-in numeric types, let's now see sequences.
Booleans
Boolean algebra is that subset of algebra in which the values of the variables are the truth values: true and false. In Python, True
and False
are two keywords that are used to represent truth values. Booleans are a subclass of integers, and behave respectively like 1 and 0. The equivalent of the int
class for Booleans is the bool
class, which returns either True
or False
. Every built-in Python object has a value in the Boolean context, which means they basically evaluate to either True
or False
when fed to the bool
function. We'll see all about this in Chapter 3, Iterating and Making Decisions.
Boolean values can be combined in Boolean expressions using the logical operators and
, or
, and not
. Again, we'll see them in full in the next chapter, so for now let's just see a simple example:
>>> int(True) # True behaves like 1 1 >>> int(False) # False behaves like 0 0 >>> bool(1) # 1 evaluates to True in a boolean context True >>> bool(-42) # and so does every non-zero number True >>> bool(0) # 0 evaluates to False False >>> # quick peak at the operators (and, or, not) >>> not True False >>> not False True >>> True and True True >>> False or True True
You can see that True
and False
are subclasses of integers when you try to add them. Python upcasts them to integers and performs addition:
>>> 1 + True 2 >>> False + 42 42 >>> 7 - True 6
Note
Upcasting is a type conversion operation that goes from a subclass to its parent. In the example presented here, True
and False
, which belong to a class derived from the integer class, are converted back to integers when needed. This topic is about inheritance and will be explained in detail in Chapter 6, Advanced Concepts – OOP, Decorators, and Iterators.
Real numbers, or floating point numbers, are represented in Python according to the IEEE 754 double-precision binary floating-point format, which is stored in 64 bits of information divided into three sections: sign, exponent, and mantissa.
Note
Quench your thirst for knowledge about this format on Wikipedia: http://en.wikipedia.org/wiki/Double-precision_floating-point_format
Usually programming languages give coders two different formats: single and double precision. The former taking up 32 bits of memory, and the latter 64. Python supports only the double format. Let's see a simple example:
>>> pi = 3.1415926536 # how many digits of PI can you remember? >>> radius = 4.5 >>> area = pi * (radius ** 2) >>> area 63.61725123519331
The sys.float_info
struct sequence holds information about how floating point numbers will behave on your system. This is what I see on my box:
>>> import sys >>> sys.float_info sys.float_info(max=1.7976931348623157e+308, max_exp=1024, max_10_exp=308, min=2.2250738585072014e-308, min_exp=-1021, min_10_exp=-307, dig=15, mant_dig=53, epsilon=2.220446049250313e-16, radix=2, rounds=1)
Let's make a few considerations here: we have 64 bits to represent float numbers. This means we can represent at most 2 ** 64 == 18,446,744,073,709,551,616
numbers with that amount of bits. Take a look at the max
and epsilon
value for the float numbers, and you'll realize it's impossible to represent them all. There is just not enough space so they are approximated to the closest representable number. You probably think that only extremely big or extremely small numbers suffer from this issue. Well, think again:
>>> 3 * 0.1 – 0.3 # this should be 0!!! 5.551115123125783e-17
What does this tell you? It tells you that double precision numbers suffer from approximation issues even when it comes to simple numbers like 0.1 or 0.3. Why is this important? It can be a big problem if you're handling prices, or financial calculations, or any kind of data that needs not to be approximated. Don't worry, Python gives you the Decimal type, which doesn't suffer from these issues, we'll see them in a bit.
Python gives you complex numbers support out of the box. If you don't know what complex numbers are, you can look them up on the Web. They are numbers that can be expressed in the form a + ib where a and b are real numbers, and i (or j if you're an engineer) is the imaginary unit, that is, the square root of -1. a and b are called respectively the real and imaginary part of the number.
It's actually unlikely you'll be using them, unless you're coding something scientific. Let's see a small example:
>>> c = 3.14 + 2.73j >>> c.real # real part 3.14 >>> c.imag # imaginary part 2.73 >>> c.conjugate() # conjugate of A + Bj is A - Bj (3.14-2.73j) >>> c * 2 # multiplication is allowed (6.28+5.46j) >>> c ** 2 # power operation as well (2.4067000000000007+17.1444j) >>> d = 1 + 1j # addition and subtraction as well >>> c - d (2.14+1.73j)
Let's finish the tour of the number department with a look at fractions and decimals. Fractions hold a rational numerator and denominator in their lowest forms. Let's see a quick example:
>>> from fractions import Fraction >>> Fraction(10, 6) # mad hatter? Fraction(5, 3) # notice it's been reduced to lowest terms >>> Fraction(1, 3) + Fraction(2, 3) # 1/3 + 2/3 = 3/3 = 1/1 Fraction(1, 1) >>> f = Fraction(10, 6) >>> f.numerator 5 >>> f.denominator 3
Although they can be very useful at times, it's not that common to spot them in commercial software. Much easier instead, is to see decimal numbers being used in all those contexts where precision is everything, for example, scientific and financial calculations.
Note
It's important to remember that arbitrary precision decimal numbers come at a price in performance, of course. The amount of data to be stored for each number is far greater than it is for fractions or floats as well as the way they are handled, which requires the Python interpreter much more work behind the scenes. Another interesting thing to know is that you can get and set the precision by accessing decimal.getcontext().prec
.
Let's see a quick example with Decimal
numbers:
>>> from decimal import Decimal as D # rename for brevity >>> D(3.14) # pi, from float, so approximation issues Decimal('3.140000000000000124344978758017532527446746826171875') >>> D('3.14') # pi, from a string, so no approximation issues Decimal('3.14') >>> D(0.1) * D(3) - D(0.3) # from float, we still have the issue Decimal('2.775557561565156540423631668E-17') >>> D('0.1') * D(3) - D('0.3') # from string, all perfect Decimal('0.0')
Notice that when we construct a Decimal
number from a float
, it takes on all the approximation issues the float
may come from. On the other hand, when the Decimal
has no approximation issues, for example, when we feed an int
or a string
representation to the constructor, then the calculation has no quirky behavior. When it comes to money, use decimals.
This concludes our introduction to built-in numeric types, let's now see sequences.
Reals
Real numbers, or floating point numbers, are represented in Python according to the IEEE 754 double-precision binary floating-point format, which is stored in 64 bits of information divided into three sections: sign, exponent, and mantissa.
Note
Quench your thirst for knowledge about this format on Wikipedia: http://en.wikipedia.org/wiki/Double-precision_floating-point_format
Usually programming languages give coders two different formats: single and double precision. The former taking up 32 bits of memory, and the latter 64. Python supports only the double format. Let's see a simple example:
>>> pi = 3.1415926536 # how many digits of PI can you remember? >>> radius = 4.5 >>> area = pi * (radius ** 2) >>> area 63.61725123519331
The sys.float_info
struct sequence holds information about how floating point numbers will behave on your system. This is what I see on my box:
>>> import sys >>> sys.float_info sys.float_info(max=1.7976931348623157e+308, max_exp=1024, max_10_exp=308, min=2.2250738585072014e-308, min_exp=-1021, min_10_exp=-307, dig=15, mant_dig=53, epsilon=2.220446049250313e-16, radix=2, rounds=1)
Let's make a few considerations here: we have 64 bits to represent float numbers. This means we can represent at most 2 ** 64 == 18,446,744,073,709,551,616
numbers with that amount of bits. Take a look at the max
and epsilon
value for the float numbers, and you'll realize it's impossible to represent them all. There is just not enough space so they are approximated to the closest representable number. You probably think that only extremely big or extremely small numbers suffer from this issue. Well, think again:
>>> 3 * 0.1 – 0.3 # this should be 0!!! 5.551115123125783e-17
What does this tell you? It tells you that double precision numbers suffer from approximation issues even when it comes to simple numbers like 0.1 or 0.3. Why is this important? It can be a big problem if you're handling prices, or financial calculations, or any kind of data that needs not to be approximated. Don't worry, Python gives you the Decimal type, which doesn't suffer from these issues, we'll see them in a bit.
Python gives you complex numbers support out of the box. If you don't know what complex numbers are, you can look them up on the Web. They are numbers that can be expressed in the form a + ib where a and b are real numbers, and i (or j if you're an engineer) is the imaginary unit, that is, the square root of -1. a and b are called respectively the real and imaginary part of the number.
It's actually unlikely you'll be using them, unless you're coding something scientific. Let's see a small example:
>>> c = 3.14 + 2.73j >>> c.real # real part 3.14 >>> c.imag # imaginary part 2.73 >>> c.conjugate() # conjugate of A + Bj is A - Bj (3.14-2.73j) >>> c * 2 # multiplication is allowed (6.28+5.46j) >>> c ** 2 # power operation as well (2.4067000000000007+17.1444j) >>> d = 1 + 1j # addition and subtraction as well >>> c - d (2.14+1.73j)
Let's finish the tour of the number department with a look at fractions and decimals. Fractions hold a rational numerator and denominator in their lowest forms. Let's see a quick example:
>>> from fractions import Fraction >>> Fraction(10, 6) # mad hatter? Fraction(5, 3) # notice it's been reduced to lowest terms >>> Fraction(1, 3) + Fraction(2, 3) # 1/3 + 2/3 = 3/3 = 1/1 Fraction(1, 1) >>> f = Fraction(10, 6) >>> f.numerator 5 >>> f.denominator 3
Although they can be very useful at times, it's not that common to spot them in commercial software. Much easier instead, is to see decimal numbers being used in all those contexts where precision is everything, for example, scientific and financial calculations.
Note
It's important to remember that arbitrary precision decimal numbers come at a price in performance, of course. The amount of data to be stored for each number is far greater than it is for fractions or floats as well as the way they are handled, which requires the Python interpreter much more work behind the scenes. Another interesting thing to know is that you can get and set the precision by accessing decimal.getcontext().prec
.
Let's see a quick example with Decimal
numbers:
>>> from decimal import Decimal as D # rename for brevity >>> D(3.14) # pi, from float, so approximation issues Decimal('3.140000000000000124344978758017532527446746826171875') >>> D('3.14') # pi, from a string, so no approximation issues Decimal('3.14') >>> D(0.1) * D(3) - D(0.3) # from float, we still have the issue Decimal('2.775557561565156540423631668E-17') >>> D('0.1') * D(3) - D('0.3') # from string, all perfect Decimal('0.0')
Notice that when we construct a Decimal
number from a float
, it takes on all the approximation issues the float
may come from. On the other hand, when the Decimal
has no approximation issues, for example, when we feed an int
or a string
representation to the constructor, then the calculation has no quirky behavior. When it comes to money, use decimals.
This concludes our introduction to built-in numeric types, let's now see sequences.
Complex numbers
Python gives you complex numbers support out of the box. If you don't know what complex numbers are, you can look them up on the Web. They are numbers that can be expressed in the form a + ib where a and b are real numbers, and i (or j if you're an engineer) is the imaginary unit, that is, the square root of -1. a and b are called respectively the real and imaginary part of the number.
It's actually unlikely you'll be using them, unless you're coding something scientific. Let's see a small example:
>>> c = 3.14 + 2.73j >>> c.real # real part 3.14 >>> c.imag # imaginary part 2.73 >>> c.conjugate() # conjugate of A + Bj is A - Bj (3.14-2.73j) >>> c * 2 # multiplication is allowed (6.28+5.46j) >>> c ** 2 # power operation as well (2.4067000000000007+17.1444j) >>> d = 1 + 1j # addition and subtraction as well >>> c - d (2.14+1.73j)
Let's finish the tour of the number department with a look at fractions and decimals. Fractions hold a rational numerator and denominator in their lowest forms. Let's see a quick example:
>>> from fractions import Fraction >>> Fraction(10, 6) # mad hatter? Fraction(5, 3) # notice it's been reduced to lowest terms >>> Fraction(1, 3) + Fraction(2, 3) # 1/3 + 2/3 = 3/3 = 1/1 Fraction(1, 1) >>> f = Fraction(10, 6) >>> f.numerator 5 >>> f.denominator 3
Although they can be very useful at times, it's not that common to spot them in commercial software. Much easier instead, is to see decimal numbers being used in all those contexts where precision is everything, for example, scientific and financial calculations.
Note
It's important to remember that arbitrary precision decimal numbers come at a price in performance, of course. The amount of data to be stored for each number is far greater than it is for fractions or floats as well as the way they are handled, which requires the Python interpreter much more work behind the scenes. Another interesting thing to know is that you can get and set the precision by accessing decimal.getcontext().prec
.
Let's see a quick example with Decimal
numbers:
>>> from decimal import Decimal as D # rename for brevity >>> D(3.14) # pi, from float, so approximation issues Decimal('3.140000000000000124344978758017532527446746826171875') >>> D('3.14') # pi, from a string, so no approximation issues Decimal('3.14') >>> D(0.1) * D(3) - D(0.3) # from float, we still have the issue Decimal('2.775557561565156540423631668E-17') >>> D('0.1') * D(3) - D('0.3') # from string, all perfect Decimal('0.0')
Notice that when we construct a Decimal
number from a float
, it takes on all the approximation issues the float
may come from. On the other hand, when the Decimal
has no approximation issues, for example, when we feed an int
or a string
representation to the constructor, then the calculation has no quirky behavior. When it comes to money, use decimals.
This concludes our introduction to built-in numeric types, let's now see sequences.
Fractions and decimals
Let's finish the tour of the number department with a look at fractions and decimals. Fractions hold a rational numerator and denominator in their lowest forms. Let's see a quick example:
>>> from fractions import Fraction >>> Fraction(10, 6) # mad hatter? Fraction(5, 3) # notice it's been reduced to lowest terms >>> Fraction(1, 3) + Fraction(2, 3) # 1/3 + 2/3 = 3/3 = 1/1 Fraction(1, 1) >>> f = Fraction(10, 6) >>> f.numerator 5 >>> f.denominator 3
Although they can be very useful at times, it's not that common to spot them in commercial software. Much easier instead, is to see decimal numbers being used in all those contexts where precision is everything, for example, scientific and financial calculations.
Note
It's important to remember that arbitrary precision decimal numbers come at a price in performance, of course. The amount of data to be stored for each number is far greater than it is for fractions or floats as well as the way they are handled, which requires the Python interpreter much more work behind the scenes. Another interesting thing to know is that you can get and set the precision by accessing decimal.getcontext().prec
.
Let's see a quick example with Decimal
numbers:
>>> from decimal import Decimal as D # rename for brevity >>> D(3.14) # pi, from float, so approximation issues Decimal('3.140000000000000124344978758017532527446746826171875') >>> D('3.14') # pi, from a string, so no approximation issues Decimal('3.14') >>> D(0.1) * D(3) - D(0.3) # from float, we still have the issue Decimal('2.775557561565156540423631668E-17') >>> D('0.1') * D(3) - D('0.3') # from string, all perfect Decimal('0.0')
Notice that when we construct a Decimal
number from a float
, it takes on all the approximation issues the float
may come from. On the other hand, when the Decimal
has no approximation issues, for example, when we feed an int
or a string
representation to the constructor, then the calculation has no quirky behavior. When it comes to money, use decimals.
This concludes our introduction to built-in numeric types, let's now see sequences.
Let's start with immutable sequences: strings, tuples, and bytes.
Textual data in Python is handled with str objects, more commonly known as strings. They are immutable sequences of unicode code points. Unicode code points can represent a character, but can also have other meanings, such as formatting data for example. Python, unlike other languages, doesn't have a char
type, so a single character is rendered simply by a string of length 1. Unicode is an excellent way to handle data, and should be used for the internals of any application. When it comes to store textual data though, or send it on the network, you may want to encode it, using an appropriate encoding for the medium you're using. String literals are written in Python using single, double or triple quotes (both single or double). If built with triple quotes, a string can span on multiple lines. An example will clarify the picture:
>>> # 4 ways to make a string >>> str1 = 'This is a string. We built it with single quotes.' >>> str2 = "This is also a string, but built with double quotes." >>> str3 = '''This is built using triple quotes, ... so it can span multiple lines.''' >>> str4 = """This too ... is a multiline one ... built with triple double-quotes.""" >>> str4 #A 'This too\nis a multiline one\nbuilt with triple double-quotes.' >>> print(str4) #B This too is a multiline one built with triple double-quotes.
In #A
and #B
, we print str4
, first implicitly, then explicitly using the print
function. A nice exercise would be to find out why they are different. Are you up to the challenge? (hint, look up the str
function)
Strings, like any sequence, have a length. You can get this by calling the len
function:
>>> len(str1) 49
Using the encode
/decode
methods, we can encode unicode strings and decode bytes objects. Utf-8 is a variable length character encoding, capable of encoding all possible unicode code points. It is the dominant encoding for the Web (and not only). Notice also that by adding a literal b
in front of a string declaration, we're creating a bytes object.
>>> s = "This is üŋíc0de" # unicode string: code points >>> type(s) <class 'str'> >>> encoded_s = s.encode('utf-8') # utf-8 encoded version of s >>> encoded_s b'This is \xc3\xbc\xc5\x8b\xc3\xadc0de' # result: bytes object >>> type(encoded_s) # another way to verify it <class 'bytes'> >>> encoded_s.decode('utf-8') # let's revert to the original 'This is üŋíc0de' >>> bytes_obj = b"A bytes object" # a bytes object >>> type(bytes_obj) <class 'bytes'>
When manipulating sequences, it's very common to have to access them at one precise position (indexing), or to get a subsequence out of them (slicing). When dealing with immutable sequences, both operations are read-only.
While indexing comes in one form, a zero-based access to any position within the sequence, slicing comes in different forms. When you get a slice of a sequence, you can specify the start
and stop
positions, and the step
. They are separated with a colon (:
) like this: my_sequence[start:stop:step]
. All the arguments are optional, start
is inclusive, stop
is exclusive. It's much easier to show an example, rather than explain them further in words:
>>> s = "The trouble is you think you have time." >>> s[0] # indexing at position 0, which is the first char 'T' >>> s[5] # indexing at position 5, which is the sixth char 'r' >>> s[:4] # slicing, we specify only the stop position 'The ' >>> s[4:] # slicing, we specify only the start position 'trouble is you think you have time.' >>> s[2:14] # slicing, both start and stop positions 'e trouble is' >>> s[2:14:3] # slicing, start, stop and step (every 3 chars) 'erb ' >>> s[:] # quick way of making a copy 'The trouble is you think you have time.'
Of all the lines, the last one is probably the most interesting. If you don't specify a parameter, Python will fill in the default for you. In this case, start
will be the start of the string, stop
will be the end of the sting, and step
will be the default 1. This is an easy and quick way of obtaining a copy of the string s
(same value, but different object). Can you find a way to get the reversed copy of a string using slicing? (don't look it up, find it for yourself)
The last immutable sequence type we're going to see is the tuple. A tuple is a sequence of arbitrary Python objects. In a tuple, items are separated by commas. They are used everywhere in Python, because they allow for patterns that are hard to reproduce in other languages. Sometimes tuples are used implicitly, for example to set up multiple variables on one line, or to allow a function to return multiple different objects (usually a function returns one object only, in many other languages), and even in the Python console, you can use tuples implicitly to print multiple elements with one single instruction. We'll see examples for all these cases:
>>> t = () # empty tuple >>> type(t) <class 'tuple'> >>> one_element_tuple = (42, ) # you need the comma! >>> three_elements_tuple = (1, 3, 5) >>> a, b, c = 1, 2, 3 # tuple for multiple assignment >>> a, b, c # implicit tuple to print with one instruction (1, 2, 3) >>> 3 in three_elements_tuple # membership test True
Notice that the membership operator in
can also be used with lists, strings, dictionaries, and in general with collection and sequence objects.
Note
Notice that to create a tuple with one item, we need to put that comma after the item. The reason is that without the comma that item is just itself wrapped in braces, kind of in a redundant mathematical expression. Notice also that on assignment, braces are optional so my_tuple = 1, 2, 3
is the same as my_tuple = (1, 2, 3)
.
One thing that tuple assignment allows us to do, is one-line swaps, with no need for a third temporary variable. Let's see first a more traditional way of doing it:
>>> a, b = 1, 2 >>> c = a # we need three lines and a temporary var c >>> a = b >>> b = c >>> a, b # a and b have been swapped (2, 1)
And now let's see how we would do it in Python:
>>> a, b = b, a # this is the Pythonic way to do it >>> a, b (1, 2)
Take a look at the line that shows you the Pythonic way of swapping two values: do you remember what I wrote in Chapter 1, Introduction and First Steps – Take a Deep Breath. A Python program is typically one-fifth to one-third the size of equivalent Java or C++ code, and features like one-line swaps contribute to this. Python is elegant, where elegance in this context means also economy.
Because they are immutable, tuples can be used as keys for dictionaries (we'll see this shortly). The dict
objects need keys to be immutable because if they could change, then the value they reference wouldn't be found any more (because the path to it depends on the key). If you are into data structures, you know how nice a feature this one is to have. To me, tuples are Python's built-in data that most closely represent a mathematical vector. This doesn't mean that this was the reason for which they were created though. Tuples usually contain an heterogeneous sequence of elements, while on the other hand lists are most of the times homogeneous. Moreover, tuples are normally accessed via unpacking or indexing, while lists are usually iterated over.
Strings and bytes
Textual data in Python is handled with str objects, more commonly known as strings. They are immutable sequences of unicode code points. Unicode code points can represent a character, but can also have other meanings, such as formatting data for example. Python, unlike other languages, doesn't have a char
type, so a single character is rendered simply by a string of length 1. Unicode is an excellent way to handle data, and should be used for the internals of any application. When it comes to store textual data though, or send it on the network, you may want to encode it, using an appropriate encoding for the medium you're using. String literals are written in Python using single, double or triple quotes (both single or double). If built with triple quotes, a string can span on multiple lines. An example will clarify the picture:
>>> # 4 ways to make a string >>> str1 = 'This is a string. We built it with single quotes.' >>> str2 = "This is also a string, but built with double quotes." >>> str3 = '''This is built using triple quotes, ... so it can span multiple lines.''' >>> str4 = """This too ... is a multiline one ... built with triple double-quotes.""" >>> str4 #A 'This too\nis a multiline one\nbuilt with triple double-quotes.' >>> print(str4) #B This too is a multiline one built with triple double-quotes.
In #A
and #B
, we print str4
, first implicitly, then explicitly using the print
function. A nice exercise would be to find out why they are different. Are you up to the challenge? (hint, look up the str
function)
Strings, like any sequence, have a length. You can get this by calling the len
function:
>>> len(str1) 49
Using the encode
/decode
methods, we can encode unicode strings and decode bytes objects. Utf-8 is a variable length character encoding, capable of encoding all possible unicode code points. It is the dominant encoding for the Web (and not only). Notice also that by adding a literal b
in front of a string declaration, we're creating a bytes object.
>>> s = "This is üŋíc0de" # unicode string: code points >>> type(s) <class 'str'> >>> encoded_s = s.encode('utf-8') # utf-8 encoded version of s >>> encoded_s b'This is \xc3\xbc\xc5\x8b\xc3\xadc0de' # result: bytes object >>> type(encoded_s) # another way to verify it <class 'bytes'> >>> encoded_s.decode('utf-8') # let's revert to the original 'This is üŋíc0de' >>> bytes_obj = b"A bytes object" # a bytes object >>> type(bytes_obj) <class 'bytes'>
When manipulating sequences, it's very common to have to access them at one precise position (indexing), or to get a subsequence out of them (slicing). When dealing with immutable sequences, both operations are read-only.
While indexing comes in one form, a zero-based access to any position within the sequence, slicing comes in different forms. When you get a slice of a sequence, you can specify the start
and stop
positions, and the step
. They are separated with a colon (:
) like this: my_sequence[start:stop:step]
. All the arguments are optional, start
is inclusive, stop
is exclusive. It's much easier to show an example, rather than explain them further in words:
>>> s = "The trouble is you think you have time." >>> s[0] # indexing at position 0, which is the first char 'T' >>> s[5] # indexing at position 5, which is the sixth char 'r' >>> s[:4] # slicing, we specify only the stop position 'The ' >>> s[4:] # slicing, we specify only the start position 'trouble is you think you have time.' >>> s[2:14] # slicing, both start and stop positions 'e trouble is' >>> s[2:14:3] # slicing, start, stop and step (every 3 chars) 'erb ' >>> s[:] # quick way of making a copy 'The trouble is you think you have time.'
Of all the lines, the last one is probably the most interesting. If you don't specify a parameter, Python will fill in the default for you. In this case, start
will be the start of the string, stop
will be the end of the sting, and step
will be the default 1. This is an easy and quick way of obtaining a copy of the string s
(same value, but different object). Can you find a way to get the reversed copy of a string using slicing? (don't look it up, find it for yourself)
The last immutable sequence type we're going to see is the tuple. A tuple is a sequence of arbitrary Python objects. In a tuple, items are separated by commas. They are used everywhere in Python, because they allow for patterns that are hard to reproduce in other languages. Sometimes tuples are used implicitly, for example to set up multiple variables on one line, or to allow a function to return multiple different objects (usually a function returns one object only, in many other languages), and even in the Python console, you can use tuples implicitly to print multiple elements with one single instruction. We'll see examples for all these cases:
>>> t = () # empty tuple >>> type(t) <class 'tuple'> >>> one_element_tuple = (42, ) # you need the comma! >>> three_elements_tuple = (1, 3, 5) >>> a, b, c = 1, 2, 3 # tuple for multiple assignment >>> a, b, c # implicit tuple to print with one instruction (1, 2, 3) >>> 3 in three_elements_tuple # membership test True
Notice that the membership operator in
can also be used with lists, strings, dictionaries, and in general with collection and sequence objects.
Note
Notice that to create a tuple with one item, we need to put that comma after the item. The reason is that without the comma that item is just itself wrapped in braces, kind of in a redundant mathematical expression. Notice also that on assignment, braces are optional so my_tuple = 1, 2, 3
is the same as my_tuple = (1, 2, 3)
.
One thing that tuple assignment allows us to do, is one-line swaps, with no need for a third temporary variable. Let's see first a more traditional way of doing it:
>>> a, b = 1, 2 >>> c = a # we need three lines and a temporary var c >>> a = b >>> b = c >>> a, b # a and b have been swapped (2, 1)
And now let's see how we would do it in Python:
>>> a, b = b, a # this is the Pythonic way to do it >>> a, b (1, 2)
Take a look at the line that shows you the Pythonic way of swapping two values: do you remember what I wrote in Chapter 1, Introduction and First Steps – Take a Deep Breath. A Python program is typically one-fifth to one-third the size of equivalent Java or C++ code, and features like one-line swaps contribute to this. Python is elegant, where elegance in this context means also economy.
Because they are immutable, tuples can be used as keys for dictionaries (we'll see this shortly). The dict
objects need keys to be immutable because if they could change, then the value they reference wouldn't be found any more (because the path to it depends on the key). If you are into data structures, you know how nice a feature this one is to have. To me, tuples are Python's built-in data that most closely represent a mathematical vector. This doesn't mean that this was the reason for which they were created though. Tuples usually contain an heterogeneous sequence of elements, while on the other hand lists are most of the times homogeneous. Moreover, tuples are normally accessed via unpacking or indexing, while lists are usually iterated over.
Encoding and decoding strings
Using the encode
/decode
methods, we can encode unicode strings and decode bytes objects. Utf-8 is a variable length character encoding, capable of encoding all possible unicode code points. It is the dominant encoding for the Web (and not only). Notice also that by adding a literal b
in front of a string declaration, we're creating a bytes object.
>>> s = "This is üŋíc0de" # unicode string: code points >>> type(s) <class 'str'> >>> encoded_s = s.encode('utf-8') # utf-8 encoded version of s >>> encoded_s b'This is \xc3\xbc\xc5\x8b\xc3\xadc0de' # result: bytes object >>> type(encoded_s) # another way to verify it <class 'bytes'> >>> encoded_s.decode('utf-8') # let's revert to the original 'This is üŋíc0de' >>> bytes_obj = b"A bytes object" # a bytes object >>> type(bytes_obj) <class 'bytes'>
When manipulating sequences, it's very common to have to access them at one precise position (indexing), or to get a subsequence out of them (slicing). When dealing with immutable sequences, both operations are read-only.
While indexing comes in one form, a zero-based access to any position within the sequence, slicing comes in different forms. When you get a slice of a sequence, you can specify the start
and stop
positions, and the step
. They are separated with a colon (:
) like this: my_sequence[start:stop:step]
. All the arguments are optional, start
is inclusive, stop
is exclusive. It's much easier to show an example, rather than explain them further in words:
>>> s = "The trouble is you think you have time." >>> s[0] # indexing at position 0, which is the first char 'T' >>> s[5] # indexing at position 5, which is the sixth char 'r' >>> s[:4] # slicing, we specify only the stop position 'The ' >>> s[4:] # slicing, we specify only the start position 'trouble is you think you have time.' >>> s[2:14] # slicing, both start and stop positions 'e trouble is' >>> s[2:14:3] # slicing, start, stop and step (every 3 chars) 'erb ' >>> s[:] # quick way of making a copy 'The trouble is you think you have time.'
Of all the lines, the last one is probably the most interesting. If you don't specify a parameter, Python will fill in the default for you. In this case, start
will be the start of the string, stop
will be the end of the sting, and step
will be the default 1. This is an easy and quick way of obtaining a copy of the string s
(same value, but different object). Can you find a way to get the reversed copy of a string using slicing? (don't look it up, find it for yourself)
The last immutable sequence type we're going to see is the tuple. A tuple is a sequence of arbitrary Python objects. In a tuple, items are separated by commas. They are used everywhere in Python, because they allow for patterns that are hard to reproduce in other languages. Sometimes tuples are used implicitly, for example to set up multiple variables on one line, or to allow a function to return multiple different objects (usually a function returns one object only, in many other languages), and even in the Python console, you can use tuples implicitly to print multiple elements with one single instruction. We'll see examples for all these cases:
>>> t = () # empty tuple >>> type(t) <class 'tuple'> >>> one_element_tuple = (42, ) # you need the comma! >>> three_elements_tuple = (1, 3, 5) >>> a, b, c = 1, 2, 3 # tuple for multiple assignment >>> a, b, c # implicit tuple to print with one instruction (1, 2, 3) >>> 3 in three_elements_tuple # membership test True
Notice that the membership operator in
can also be used with lists, strings, dictionaries, and in general with collection and sequence objects.
Note
Notice that to create a tuple with one item, we need to put that comma after the item. The reason is that without the comma that item is just itself wrapped in braces, kind of in a redundant mathematical expression. Notice also that on assignment, braces are optional so my_tuple = 1, 2, 3
is the same as my_tuple = (1, 2, 3)
.
One thing that tuple assignment allows us to do, is one-line swaps, with no need for a third temporary variable. Let's see first a more traditional way of doing it:
>>> a, b = 1, 2 >>> c = a # we need three lines and a temporary var c >>> a = b >>> b = c >>> a, b # a and b have been swapped (2, 1)
And now let's see how we would do it in Python:
>>> a, b = b, a # this is the Pythonic way to do it >>> a, b (1, 2)
Take a look at the line that shows you the Pythonic way of swapping two values: do you remember what I wrote in Chapter 1, Introduction and First Steps – Take a Deep Breath. A Python program is typically one-fifth to one-third the size of equivalent Java or C++ code, and features like one-line swaps contribute to this. Python is elegant, where elegance in this context means also economy.
Because they are immutable, tuples can be used as keys for dictionaries (we'll see this shortly). The dict
objects need keys to be immutable because if they could change, then the value they reference wouldn't be found any more (because the path to it depends on the key). If you are into data structures, you know how nice a feature this one is to have. To me, tuples are Python's built-in data that most closely represent a mathematical vector. This doesn't mean that this was the reason for which they were created though. Tuples usually contain an heterogeneous sequence of elements, while on the other hand lists are most of the times homogeneous. Moreover, tuples are normally accessed via unpacking or indexing, while lists are usually iterated over.
Indexing and slicing strings
When manipulating sequences, it's very common to have to access them at one precise position (indexing), or to get a subsequence out of them (slicing). When dealing with immutable sequences, both operations are read-only.
While indexing comes in one form, a zero-based access to any position within the sequence, slicing comes in different forms. When you get a slice of a sequence, you can specify the start
and stop
positions, and the step
. They are separated with a colon (:
) like this: my_sequence[start:stop:step]
. All the arguments are optional, start
is inclusive, stop
is exclusive. It's much easier to show an example, rather than explain them further in words:
>>> s = "The trouble is you think you have time." >>> s[0] # indexing at position 0, which is the first char 'T' >>> s[5] # indexing at position 5, which is the sixth char 'r' >>> s[:4] # slicing, we specify only the stop position 'The ' >>> s[4:] # slicing, we specify only the start position 'trouble is you think you have time.' >>> s[2:14] # slicing, both start and stop positions 'e trouble is' >>> s[2:14:3] # slicing, start, stop and step (every 3 chars) 'erb ' >>> s[:] # quick way of making a copy 'The trouble is you think you have time.'
Of all the lines, the last one is probably the most interesting. If you don't specify a parameter, Python will fill in the default for you. In this case, start
will be the start of the string, stop
will be the end of the sting, and step
will be the default 1. This is an easy and quick way of obtaining a copy of the string s
(same value, but different object). Can you find a way to get the reversed copy of a string using slicing? (don't look it up, find it for yourself)
The last immutable sequence type we're going to see is the tuple. A tuple is a sequence of arbitrary Python objects. In a tuple, items are separated by commas. They are used everywhere in Python, because they allow for patterns that are hard to reproduce in other languages. Sometimes tuples are used implicitly, for example to set up multiple variables on one line, or to allow a function to return multiple different objects (usually a function returns one object only, in many other languages), and even in the Python console, you can use tuples implicitly to print multiple elements with one single instruction. We'll see examples for all these cases:
>>> t = () # empty tuple >>> type(t) <class 'tuple'> >>> one_element_tuple = (42, ) # you need the comma! >>> three_elements_tuple = (1, 3, 5) >>> a, b, c = 1, 2, 3 # tuple for multiple assignment >>> a, b, c # implicit tuple to print with one instruction (1, 2, 3) >>> 3 in three_elements_tuple # membership test True
Notice that the membership operator in
can also be used with lists, strings, dictionaries, and in general with collection and sequence objects.
Note
Notice that to create a tuple with one item, we need to put that comma after the item. The reason is that without the comma that item is just itself wrapped in braces, kind of in a redundant mathematical expression. Notice also that on assignment, braces are optional so my_tuple = 1, 2, 3
is the same as my_tuple = (1, 2, 3)
.
One thing that tuple assignment allows us to do, is one-line swaps, with no need for a third temporary variable. Let's see first a more traditional way of doing it:
>>> a, b = 1, 2 >>> c = a # we need three lines and a temporary var c >>> a = b >>> b = c >>> a, b # a and b have been swapped (2, 1)
And now let's see how we would do it in Python:
>>> a, b = b, a # this is the Pythonic way to do it >>> a, b (1, 2)
Take a look at the line that shows you the Pythonic way of swapping two values: do you remember what I wrote in Chapter 1, Introduction and First Steps – Take a Deep Breath. A Python program is typically one-fifth to one-third the size of equivalent Java or C++ code, and features like one-line swaps contribute to this. Python is elegant, where elegance in this context means also economy.
Because they are immutable, tuples can be used as keys for dictionaries (we'll see this shortly). The dict
objects need keys to be immutable because if they could change, then the value they reference wouldn't be found any more (because the path to it depends on the key). If you are into data structures, you know how nice a feature this one is to have. To me, tuples are Python's built-in data that most closely represent a mathematical vector. This doesn't mean that this was the reason for which they were created though. Tuples usually contain an heterogeneous sequence of elements, while on the other hand lists are most of the times homogeneous. Moreover, tuples are normally accessed via unpacking or indexing, while lists are usually iterated over.
Tuples
The last immutable sequence type we're going to see is the tuple. A tuple is a sequence of arbitrary Python objects. In a tuple, items are separated by commas. They are used everywhere in Python, because they allow for patterns that are hard to reproduce in other languages. Sometimes tuples are used implicitly, for example to set up multiple variables on one line, or to allow a function to return multiple different objects (usually a function returns one object only, in many other languages), and even in the Python console, you can use tuples implicitly to print multiple elements with one single instruction. We'll see examples for all these cases:
>>> t = () # empty tuple >>> type(t) <class 'tuple'> >>> one_element_tuple = (42, ) # you need the comma! >>> three_elements_tuple = (1, 3, 5) >>> a, b, c = 1, 2, 3 # tuple for multiple assignment >>> a, b, c # implicit tuple to print with one instruction (1, 2, 3) >>> 3 in three_elements_tuple # membership test True
Notice that the membership operator in
can also be used with lists, strings, dictionaries, and in general with collection and sequence objects.
Note
Notice that to create a tuple with one item, we need to put that comma after the item. The reason is that without the comma that item is just itself wrapped in braces, kind of in a redundant mathematical expression. Notice also that on assignment, braces are optional so my_tuple = 1, 2, 3
is the same as my_tuple = (1, 2, 3)
.
One thing that tuple assignment allows us to do, is one-line swaps, with no need for a third temporary variable. Let's see first a more traditional way of doing it:
>>> a, b = 1, 2 >>> c = a # we need three lines and a temporary var c >>> a = b >>> b = c >>> a, b # a and b have been swapped (2, 1)
And now let's see how we would do it in Python:
>>> a, b = b, a # this is the Pythonic way to do it >>> a, b (1, 2)
Take a look at the line that shows you the Pythonic way of swapping two values: do you remember what I wrote in Chapter 1, Introduction and First Steps – Take a Deep Breath. A Python program is typically one-fifth to one-third the size of equivalent Java or C++ code, and features like one-line swaps contribute to this. Python is elegant, where elegance in this context means also economy.
Because they are immutable, tuples can be used as keys for dictionaries (we'll see this shortly). The dict
objects need keys to be immutable because if they could change, then the value they reference wouldn't be found any more (because the path to it depends on the key). If you are into data structures, you know how nice a feature this one is to have. To me, tuples are Python's built-in data that most closely represent a mathematical vector. This doesn't mean that this was the reason for which they were created though. Tuples usually contain an heterogeneous sequence of elements, while on the other hand lists are most of the times homogeneous. Moreover, tuples are normally accessed via unpacking or indexing, while lists are usually iterated over.
Mutable sequences differ from their immutable sisters in that they can be changed after creation. There are two mutable sequence types in Python: lists and byte arrays. I said before that the dictionary is the king of data structures in Python. I guess this makes the list its rightful queen.
Python lists are mutable sequences. They are very similar to tuples, but they don't have the restrictions due to immutability. Lists are commonly used to store collections of homogeneous objects, but there is nothing preventing you to store heterogeneous collections as well. Lists can be created in many different ways, let's see an example:
>>> [] # empty list [] >>> list() # same as [] [] >>> [1, 2, 3] # as with tuples, items are comma separated [1, 2, 3] >>> [x + 5 for x in [2, 3, 4]] # Python is magic [7, 8, 9] >>> list((1, 3, 5, 7, 9)) # list from a tuple [1, 3, 5, 7, 9] >>> list('hello') # list from a string ['h', 'e', 'l', 'l', 'o']
In the previous example, I showed you how to create a list using different techniques. I would like you to take a good look at the line that says Python is magic
, which I am not expecting you to fully understand at this point (unless you cheated and you're not a novice!). That is called a list comprehension, a very powerful functional feature of Python, which we'll see in detail in Chapter 5, Saving Time and Memory. I just wanted to make your mouth water at this point.
Creating lists is good, but the real fun comes when we use them, so let's see the main methods they gift us with:
>>> a = [1, 2, 1, 3] >>> a.append(13) # we can append anything at the end >>> a [1, 2, 1, 3, 13] >>> a.count(1) # how many `1` are there in the list? 2 >>> a.extend([5, 7]) # extend the list by another (or sequence) >>> a [1, 2, 1, 3, 13, 5, 7] >>> a.index(13) # position of `13` in the list (0-based indexing) 4 >>> a.insert(0, 17) # insert `17` at position 0 >>> a [17, 1, 2, 1, 3, 13, 5, 7] >>> a.pop() # pop (remove and return) last element 7 >>> a.pop(3) # pop element at position 3 1 >>> a [17, 1, 2, 3, 13, 5] >>> a.remove(17) # remove `17` from the list >>> a [1, 2, 3, 13, 5] >>> a.reverse() # reverse the order of the elements in the list >>> a [5, 13, 3, 2, 1] >>> a.sort() # sort the list >>> a [1, 2, 3, 5, 13] >>> a.clear() # remove all elements from the list >>> a []
The preceding code gives you a roundup of list's main methods. I want to show you how powerful they are, using extend
as an example. You can extend lists using any sequence type:
>>> a = list('hello') # makes a list from a string >>> a ['h', 'e', 'l', 'l', 'o'] >>> a.append(100) # append 100, heterogeneous type >>> a ['h', 'e', 'l', 'l', 'o', 100] >>> a.extend((1, 2, 3)) # extend using tuple >>> a ['h', 'e', 'l', 'l', 'o', 100, 1, 2, 3] >>> a.extend('...') # extend using string >>> a ['h', 'e', 'l', 'l', 'o', 100, 1, 2, 3, '.', '.', '.']
Now, let's see what are the most common operations you can do with lists:
>>> a = [1, 3, 5, 7] >>> min(a) # minimum value in the list 1 >>> max(a) # maximum value in the list 7 >>> sum(a) # sum of all values in the list 16 >>> len(a) # number of elements in the list 4 >>> b = [6, 7, 8] >>> a + b # `+` with list means concatenation [1, 3, 5, 7, 6, 7, 8] >>> a * 2 # `*` has also a special meaning [1, 3, 5, 7, 1, 3, 5, 7]
The last two lines in the preceding code are quite interesting because they introduce us to a concept called operator overloading. In short, it means that operators such as +
, -
. *
, %
, and so on, may represent different operations according to the context they are used in. It doesn't make any sense to sum two lists, right? Therefore, the +
sign is used to concatenate them. Hence, the *
sign is used to concatenate the list to itself according to the right operand. Now, let's take a step further down the rabbit hole and see something a little more interesting. I want to show you how powerful the sort method can be and how easy it is in Python to achieve results that require a great deal of effort in other languages:
>>> from operator import itemgetter >>> a = [(5, 3), (1, 3), (1, 2), (2, -1), (4, 9)] >>> sorted(a) [(1, 2), (1, 3), (2, -1), (4, 9), (5, 3)] >>> sorted(a, key=itemgetter(0)) [(1, 3), (1, 2), (2, -1), (4, 9), (5, 3)] >>> sorted(a, key=itemgetter(0, 1)) [(1, 2), (1, 3), (2, -1), (4, 9), (5, 3)] >>> sorted(a, key=itemgetter(1)) [(2, -1), (1, 2), (5, 3), (1, 3), (4, 9)] >>> sorted(a, key=itemgetter(1), reverse=True) [(4, 9), (5, 3), (1, 3), (1, 2), (2, -1)]
The preceding code deserves a little explanation. First of all, a
is a list of tuples. This means each element in a
is a tuple (a 2-tuple, to be picky). When we call sorted(some_list)
, we get a sorted version of some_list
. In this case, the sorting on a 2-tuple works by sorting them on the first item in the tuple, and on the second when the first one is the same. You can see this behavior in the result of sorted(a)
, which yields [(1, 2), (1, 3), ...]
. Python also gives us the ability to control on which element(s) of the tuple the sorting must be run against. Notice that when we instruct the sorted
function to work on the first element of each tuple (by key=itemgetter(0)
), the result is different: [(1, 3), (1, 2), ...]
. The sorting is done only on the first element of each tuple (which is the one at position 0). If we want to replicate the default behavior of a simple sorted(a)
call, we need to use key=itemgetter(0, 1)
, which tells Python to sort first on the elements at position 0 within the tuples, and then on those at position 1. Compare the results and you'll see they match.
For completeness, I included an example of sorting only on the elements at position 1, and the same but in reverse order. If you have ever seen sorting in Java, I expect you to be on your knees crying with joy at this very moment.
The Python sorting algorithm is very powerful, and it was written by Tim Peters (we've already seen this name, can you recall when?). It is aptly named Timsort, and it is a blend between merge and insertion sort and has better time performances than most other algorithms used for mainstream programming languages. Timsort is a stable sorting algorithm, which means that when multiple records have the same key, their original order is preserved. We've seen this in the result of sorted(a, key=itemgetter(0))
which has yielded [(1, 3), (1, 2), ...]
in which the order of those two tuples has been preserved because they have the same value at position 0.
To conclude our overview of mutable sequence types, let's spend a couple of minutes on the bytearray
type. Basically, they represent the mutable version of bytes
objects. They expose most of the usual methods of mutable sequences as well as most of the methods of the bytes
type. Items are integers in the range [0, 256).
Note
When it comes to intervals, I'm going to use the standard notation for open/closed ranges. A square bracket on one end means that the value is included, while a round brace means it's excluded. The granularity is usually inferred by the type of the edge elements so, for example, the interval [3, 7] means all integers between 3 and 7, inclusive. On the other hand, (3, 7) means all integers between 3 and 7 exclusive (hence 4, 5, and 6). Items in a bytearray
type are integers between 0 and 256, 0 is included, 256 is not. One reason intervals are often expressed like this is to ease coding. If we break a range [a, b) into N consecutive ranges, we can easily represent the original one as a concatenation like this:

The middle points (k i) being excluded on one end, and included on the other end, allow for easy concatenation and splitting when intervals are handled in the code.
Let's see a quick example with the type bytearray
:
>>> bytearray() # empty bytearray object bytearray(b'') >>> bytearray(10) # zero-filled instance with given length bytearray(b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00') >>> bytearray(range(5)) # bytearray from iterable of integers bytearray(b'\x00\x01\x02\x03\x04') >>> name = bytearray(b'Lina') # A - bytearray from bytes >>> name.replace(b'L', b'l') bytearray(b'lina') >>> name.endswith(b'na') True >>> name.upper() bytearray(b'LINA') >>> name.count(b'L') 1
As you can see in the preceding code, there are a few ways to create a bytearray
object. They can be useful in many situations, for example, when receiving data through a socket, they eliminate the need to concatenate data while polling, hence they prove very handy. On the line #A
, I created the name
bytearray
from the string b'Lina'
to show you how the bytearray
object exposes methods from both sequences and strings, which is extremely handy. If you think about it, they can be considered as mutable strings.
Lists
Python lists are mutable sequences. They are very similar to tuples, but they don't have the restrictions due to immutability. Lists are commonly used to store collections of homogeneous objects, but there is nothing preventing you to store heterogeneous collections as well. Lists can be created in many different ways, let's see an example:
>>> [] # empty list [] >>> list() # same as [] [] >>> [1, 2, 3] # as with tuples, items are comma separated [1, 2, 3] >>> [x + 5 for x in [2, 3, 4]] # Python is magic [7, 8, 9] >>> list((1, 3, 5, 7, 9)) # list from a tuple [1, 3, 5, 7, 9] >>> list('hello') # list from a string ['h', 'e', 'l', 'l', 'o']
In the previous example, I showed you how to create a list using different techniques. I would like you to take a good look at the line that says Python is magic
, which I am not expecting you to fully understand at this point (unless you cheated and you're not a novice!). That is called a list comprehension, a very powerful functional feature of Python, which we'll see in detail in Chapter 5, Saving Time and Memory. I just wanted to make your mouth water at this point.
Creating lists is good, but the real fun comes when we use them, so let's see the main methods they gift us with:
>>> a = [1, 2, 1, 3] >>> a.append(13) # we can append anything at the end >>> a [1, 2, 1, 3, 13] >>> a.count(1) # how many `1` are there in the list? 2 >>> a.extend([5, 7]) # extend the list by another (or sequence) >>> a [1, 2, 1, 3, 13, 5, 7] >>> a.index(13) # position of `13` in the list (0-based indexing) 4 >>> a.insert(0, 17) # insert `17` at position 0 >>> a [17, 1, 2, 1, 3, 13, 5, 7] >>> a.pop() # pop (remove and return) last element 7 >>> a.pop(3) # pop element at position 3 1 >>> a [17, 1, 2, 3, 13, 5] >>> a.remove(17) # remove `17` from the list >>> a [1, 2, 3, 13, 5] >>> a.reverse() # reverse the order of the elements in the list >>> a [5, 13, 3, 2, 1] >>> a.sort() # sort the list >>> a [1, 2, 3, 5, 13] >>> a.clear() # remove all elements from the list >>> a []
The preceding code gives you a roundup of list's main methods. I want to show you how powerful they are, using extend
as an example. You can extend lists using any sequence type:
>>> a = list('hello') # makes a list from a string >>> a ['h', 'e', 'l', 'l', 'o'] >>> a.append(100) # append 100, heterogeneous type >>> a ['h', 'e', 'l', 'l', 'o', 100] >>> a.extend((1, 2, 3)) # extend using tuple >>> a ['h', 'e', 'l', 'l', 'o', 100, 1, 2, 3] >>> a.extend('...') # extend using string >>> a ['h', 'e', 'l', 'l', 'o', 100, 1, 2, 3, '.', '.', '.']
Now, let's see what are the most common operations you can do with lists:
>>> a = [1, 3, 5, 7] >>> min(a) # minimum value in the list 1 >>> max(a) # maximum value in the list 7 >>> sum(a) # sum of all values in the list 16 >>> len(a) # number of elements in the list 4 >>> b = [6, 7, 8] >>> a + b # `+` with list means concatenation [1, 3, 5, 7, 6, 7, 8] >>> a * 2 # `*` has also a special meaning [1, 3, 5, 7, 1, 3, 5, 7]
The last two lines in the preceding code are quite interesting because they introduce us to a concept called operator overloading. In short, it means that operators such as +
, -
. *
, %
, and so on, may represent different operations according to the context they are used in. It doesn't make any sense to sum two lists, right? Therefore, the +
sign is used to concatenate them. Hence, the *
sign is used to concatenate the list to itself according to the right operand. Now, let's take a step further down the rabbit hole and see something a little more interesting. I want to show you how powerful the sort method can be and how easy it is in Python to achieve results that require a great deal of effort in other languages:
>>> from operator import itemgetter >>> a = [(5, 3), (1, 3), (1, 2), (2, -1), (4, 9)] >>> sorted(a) [(1, 2), (1, 3), (2, -1), (4, 9), (5, 3)] >>> sorted(a, key=itemgetter(0)) [(1, 3), (1, 2), (2, -1), (4, 9), (5, 3)] >>> sorted(a, key=itemgetter(0, 1)) [(1, 2), (1, 3), (2, -1), (4, 9), (5, 3)] >>> sorted(a, key=itemgetter(1)) [(2, -1), (1, 2), (5, 3), (1, 3), (4, 9)] >>> sorted(a, key=itemgetter(1), reverse=True) [(4, 9), (5, 3), (1, 3), (1, 2), (2, -1)]
The preceding code deserves a little explanation. First of all, a
is a list of tuples. This means each element in a
is a tuple (a 2-tuple, to be picky). When we call sorted(some_list)
, we get a sorted version of some_list
. In this case, the sorting on a 2-tuple works by sorting them on the first item in the tuple, and on the second when the first one is the same. You can see this behavior in the result of sorted(a)
, which yields [(1, 2), (1, 3), ...]
. Python also gives us the ability to control on which element(s) of the tuple the sorting must be run against. Notice that when we instruct the sorted
function to work on the first element of each tuple (by key=itemgetter(0)
), the result is different: [(1, 3), (1, 2), ...]
. The sorting is done only on the first element of each tuple (which is the one at position 0). If we want to replicate the default behavior of a simple sorted(a)
call, we need to use key=itemgetter(0, 1)
, which tells Python to sort first on the elements at position 0 within the tuples, and then on those at position 1. Compare the results and you'll see they match.
For completeness, I included an example of sorting only on the elements at position 1, and the same but in reverse order. If you have ever seen sorting in Java, I expect you to be on your knees crying with joy at this very moment.
The Python sorting algorithm is very powerful, and it was written by Tim Peters (we've already seen this name, can you recall when?). It is aptly named Timsort, and it is a blend between merge and insertion sort and has better time performances than most other algorithms used for mainstream programming languages. Timsort is a stable sorting algorithm, which means that when multiple records have the same key, their original order is preserved. We've seen this in the result of sorted(a, key=itemgetter(0))
which has yielded [(1, 3), (1, 2), ...]
in which the order of those two tuples has been preserved because they have the same value at position 0.
To conclude our overview of mutable sequence types, let's spend a couple of minutes on the bytearray
type. Basically, they represent the mutable version of bytes
objects. They expose most of the usual methods of mutable sequences as well as most of the methods of the bytes
type. Items are integers in the range [0, 256).
Note
When it comes to intervals, I'm going to use the standard notation for open/closed ranges. A square bracket on one end means that the value is included, while a round brace means it's excluded. The granularity is usually inferred by the type of the edge elements so, for example, the interval [3, 7] means all integers between 3 and 7, inclusive. On the other hand, (3, 7) means all integers between 3 and 7 exclusive (hence 4, 5, and 6). Items in a bytearray
type are integers between 0 and 256, 0 is included, 256 is not. One reason intervals are often expressed like this is to ease coding. If we break a range [a, b) into N consecutive ranges, we can easily represent the original one as a concatenation like this:

The middle points (k i) being excluded on one end, and included on the other end, allow for easy concatenation and splitting when intervals are handled in the code.
Let's see a quick example with the type bytearray
:
>>> bytearray() # empty bytearray object bytearray(b'') >>> bytearray(10) # zero-filled instance with given length bytearray(b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00') >>> bytearray(range(5)) # bytearray from iterable of integers bytearray(b'\x00\x01\x02\x03\x04') >>> name = bytearray(b'Lina') # A - bytearray from bytes >>> name.replace(b'L', b'l') bytearray(b'lina') >>> name.endswith(b'na') True >>> name.upper() bytearray(b'LINA') >>> name.count(b'L') 1
As you can see in the preceding code, there are a few ways to create a bytearray
object. They can be useful in many situations, for example, when receiving data through a socket, they eliminate the need to concatenate data while polling, hence they prove very handy. On the line #A
, I created the name
bytearray
from the string b'Lina'
to show you how the bytearray
object exposes methods from both sequences and strings, which is extremely handy. If you think about it, they can be considered as mutable strings.
Byte arrays
To conclude our overview of mutable sequence types, let's spend a couple of minutes on the bytearray
type. Basically, they represent the mutable version of bytes
objects. They expose most of the usual methods of mutable sequences as well as most of the methods of the bytes
type. Items are integers in the range [0, 256).
Note
When it comes to intervals, I'm going to use the standard notation for open/closed ranges. A square bracket on one end means that the value is included, while a round brace means it's excluded. The granularity is usually inferred by the type of the edge elements so, for example, the interval [3, 7] means all integers between 3 and 7, inclusive. On the other hand, (3, 7) means all integers between 3 and 7 exclusive (hence 4, 5, and 6). Items in a bytearray
type are integers between 0 and 256, 0 is included, 256 is not. One reason intervals are often expressed like this is to ease coding. If we break a range [a, b) into N consecutive ranges, we can easily represent the original one as a concatenation like this:

The middle points (k i) being excluded on one end, and included on the other end, allow for easy concatenation and splitting when intervals are handled in the code.
Let's see a quick example with the type bytearray
:
>>> bytearray() # empty bytearray object bytearray(b'') >>> bytearray(10) # zero-filled instance with given length bytearray(b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00') >>> bytearray(range(5)) # bytearray from iterable of integers bytearray(b'\x00\x01\x02\x03\x04') >>> name = bytearray(b'Lina') # A - bytearray from bytes >>> name.replace(b'L', b'l') bytearray(b'lina') >>> name.endswith(b'na') True >>> name.upper() bytearray(b'LINA') >>> name.count(b'L') 1
As you can see in the preceding code, there are a few ways to create a bytearray
object. They can be useful in many situations, for example, when receiving data through a socket, they eliminate the need to concatenate data while polling, hence they prove very handy. On the line #A
, I created the name
bytearray
from the string b'Lina'
to show you how the bytearray
object exposes methods from both sequences and strings, which is extremely handy. If you think about it, they can be considered as mutable strings.
Python also provides two set types, set
and frozenset
. The set
type is mutable, while frozenset
is immutable. They are unordered collections of immutable objects.
Hashability is a characteristic that allows an object to be used as a set member as well as a key for a dictionary, as we'll see very soon.
Objects that compare equally must have the same hash value. Sets are very commonly used to test for membership, so let's introduce the in
operator in the following example:
>>> small_primes = set() # empty set >>> small_primes.add(2) # adding one element at a time >>> small_primes.add(3) >>> small_primes.add(5) >>> small_primes {2, 3, 5} >>> small_primes.add(1) # Look what I've done, 1 is not a prime! >>> small_primes {1, 2, 3, 5} >>> small_primes.remove(1) # so let's remove it >>> 3 in small_primes # membership test True >>> 4 in small_primes False >>> 4 not in small_primes # negated membership test True >>> small_primes.add(3) # trying to add 3 again >>> small_primes {2, 3, 5} # no change, duplication is not allowed >>> bigger_primes = set([5, 7, 11, 13]) # faster creation >>> small_primes | bigger_primes # union operator `|` {2, 3, 5, 7, 11, 13} >>> small_primes & bigger_primes # intersection operator `&` {5} >>> small_primes - bigger_primes # difference operator `-` {2, 3}
In the preceding code, you can see two different ways to create a set. One creates an empty set and then adds elements one at a time. The other creates the set using a list of numbers as argument to the constructor, which does all the work for us. Of course, you can create a set from a list or tuple (or any iterable) and then you can add and remove members from the set as you please.
Another way of creating a set is by simply using the curly braces notation, like this:
>>> small_primes = {2, 3, 5, 5, 3} >>> small_primes {2, 3, 5}
Notice I added some duplication to emphasize that the result set won't have any.
Let's see an example about the immutable counterpart of the set type: frozenset
.
>>> small_primes = frozenset([2, 3, 5, 7]) >>> bigger_primes = frozenset([5, 7, 11]) >>> small_primes.add(11) # we cannot add to a frozenset Traceback (most recent call last): File "<stdin>", line 1, in <module> AttributeError: 'frozenset' object has no attribute 'add' >>> small_primes.remove(2) # neither we can remove Traceback (most recent call last): File "<stdin>", line 1, in <module> AttributeError: 'frozenset' object has no attribute 'remove' >>> small_primes & bigger_primes # intersect, union, etc. allowed frozenset({5, 7})
As you can see, frozenset
objects are quite limited in respect of their mutable counterpart. They still prove very effective for membership test, union, intersection and difference operations, and for performance reasons.
Of all the built-in Python data types, the dictionary is probably the most interesting one. It's the only standard mapping type, and it is the backbone of every Python object.
A dictionary maps keys to values. Keys need to be hashable objects, while values can be of any arbitrary type. Dictionaries are mutable objects.
There are quite a few different ways to create a dictionary, so let me give you a simple example of how to create a dictionary equal to {'A': 1, 'Z': -1}
in five different ways:
>>> a = dict(A=1, Z=-1) >>> b = {'A': 1, 'Z': -1} >>> c = dict(zip(['A', 'Z'], [1, -1])) >>> d = dict([('A', 1), ('Z', -1)]) >>> e = dict({'Z': -1, 'A': 1}) >>> a == b == c == d == e # are they all the same? True # indeed!
Have you noticed those double equals? Assignment is done with one equal, while to check whether an object is the same as another one (or 5 in one go, in this case), we use double equals. There is also another way to compare objects, which involves the is operator, and checks whether the two objects are the same (if they have the same ID, not just the value), but unless you have a good reason to use it, you should use the double equal instead. In the preceding code, I also used one nice function: zip
. It is named after the real-life zip, which glues together two things taking one element from each at a time. Let me show you an example:
>>> list(zip(['h', 'e', 'l', 'l', 'o'], [1, 2, 3, 4, 5])) [('h', 1), ('e', 2), ('l', 3), ('l', 4), ('o', 5)] >>> list(zip('hello', range(1, 6))) # equivalent, more Pythonic [('h', 1), ('e', 2), ('l', 3), ('l', 4), ('o', 5)]
In the preceding example, I have created the same list in two different ways, one more explicit, and the other a little bit more Pythonic. Forget for a moment that I had to wrap the list
constructor around the zip
call (the reason is because zip
returns an iterator, not a list
), and concentrate on the result. See how zip
has coupled the first elements of its two arguments together, then the second ones, then the third ones, and so on and so forth? Take a look at your pants (or at your purse if you're a lady) and you'll see the same behavior in your actual zip. But let's go back to dictionaries and see how many wonderful methods they expose for allowing us to manipulate them as we want. Let's start with the basic operations:
>>> d = {} >>> d['a'] = 1 # let's set a couple of (key, value) pairs >>> d['b'] = 2 >>> len(d) # how many pairs? 2 >>> d['a'] # what is the value of 'a'? 1 >>> d # how does `d` look now? {'a': 1, 'b': 2} >>> del d['a'] # let's remove `a` >>> d {'b': 2} >>> d['c'] = 3 # let's add 'c': 3 >>> 'c' in d # membership is checked against the keys True >>> 3 in d # not the values False >>> 'e' in d False >>> d.clear() # let's clean everything from this dictionary >>> d {}
Notice how accessing keys of a dictionary, regardless of the type of operation we're performing, is done through square brackets. Do you remember strings, list, and tuples? We were accessing elements at some position through square brackets as well. Yet another example of Python's consistency.
Let's see now three special objects called dictionary views: keys
, values
, and items
. These objects provide a dynamic view of the dictionary entries and they change when the dictionary changes. keys()
returns all the keys in the dictionary, values()
returns all the values in the dictionary, and items()
returns all the (key, value) pairs in the dictionary.
Note
It's very important to know that, even if a dictionary is not intrinsically ordered, according to the Python documentation: "Keys and values are iterated over in an arbitrary order which is non-random, varies across Python implementations, and depends on the dictionary's history of insertions and deletions. If keys, values and items views are iterated over with no intervening modifications to the dictionary, the order of items will directly correspond."
Enough with this chatter, let's put all this down into code:
>>> d = dict(zip('hello', range(5))) >>> d {'e': 1, 'h': 0, 'o': 4, 'l': 3} >>> d.keys() dict_keys(['e', 'h', 'o', 'l']) >>> d.values() dict_values([1, 0, 4, 3]) >>> d.items() dict_items([('e', 1), ('h', 0), ('o', 4), ('l', 3)]) >>> 3 in d.values() True >>> ('o', 4) in d.items() True
A few things to notice in the preceding code. First, notice how we're creating a dictionary by iterating over the zipped version of the string 'hello'
and the list [0, 1, 2, 3, 4]
. The string 'hello'
has two 'l'
characters inside, and they are paired up with the values 2 and 3 by the zip function. Notice how in the dictionary, the second occurrence of the 'l'
key (the one with value 3), overwrites the first one (the one with value 2). Another thing to notice is that when asking for any view, the original order is lost, but is consistent within the views, as expected. Notice also that you may have different results when you try this code on your machine. Python doesn't guarantee that, it only guarantees the consistency of the order in which the views are presented.
We'll see how these views are fundamental tools when we talk about iterating over collections. Let's take a look now at some other methods exposed by Python's dictionaries, there's plenty of them and they are very useful:
>>> d {'e': 1, 'h': 0, 'o': 4, 'l': 3} >>> d.popitem() # removes a random item ('e', 1) >>> d {'h': 0, 'o': 4, 'l': 3} >>> d.pop('l') # remove item with key `l` 3 >>> d.pop('not-a-key') # remove a key not in dictionary: KeyError Traceback (most recent call last): File "<stdin>", line 1, in <module> KeyError: 'not-a-key' >>> d.pop('not-a-key', 'default-value') # with a default value? 'default-value' # we get the default value >>> d.update({'another': 'value'}) # we can update dict this way >>> d.update(a=13) # or this way (like a function call) >>> d {'a': 13, 'another': 'value', 'h': 0, 'o': 4} >>> d.get('a') # same as d['a'] but if key is missing no KeyError 13 >>> d.get('a', 177) # default value used if key is missing 13 >>> d.get('b', 177) # like in this case 177 >>> d.get('b') # key is not there, so None is returned
All these methods are quite simple to understand, but it's worth talking about that None
, for a moment. Every function in Python returns None
, unless the return
statement is explicitly used, but we'll see this when we explore functions. None
is frequently used to represent the absence of a value, as when default arguments are not passed to a function. Some inexperienced coders sometimes write code that returns either False
or None
. Both False
and None
evaluate to False
so it may seem there is not much difference between them. But actually, I would argue there is quite an important difference: False
means that we have information, and the information we have is False
. None
means no information. And no information is very different from an information, which is False
. In layman's terms, if you ask your mechanic "is my car ready?" there is a big difference between the answer "No, it's not" (False) and "I have no idea" (None).
One last method I really like of dictionaries is setdefault
. It behaves like get
, but also sets the key with the given value if it is not there. Let's see and example:
>>> d = {} >>> d.setdefault('a', 1) # 'a' is missing, we get default value 1 >>> d {'a': 1} # also, the key/value pair ('a', 1) has now been added >>> d.setdefault('a', 5) # let's try to override the value 1 >>> d {'a': 1} # didn't work, as expected
So, we're now at the end of this tour. Test your knowledge about dictionaries trying to foresee how d
looks like after this line.
>>> d = {} >>> d.setdefault('a', {}).setdefault('b', []).append(1)
It's not that complicated, but don't worry if you don't get it immediately. I just wanted to spur you to experiment with dictionaries.
This concludes our tour of built-in data types. Before I make some considerations about what we've seen in this chapter, I want to briefly take a peek at the collections
module.
When Python general purpose built-in containers (tuple
, list
, set
, and dict
) aren't enough, we can find specialized container data types in the collections
module. They are:
We don't have the room to cover all of them, but you can find plenty of examples in the official documentation, so here I'll just give a small example to show you namedtuple
, defaultdict
, and ChainMap
.
A namedtuple
is a tuple
-like object that has fields accessible by attribute lookup as well as being indexable and iterable (it's actually a subclass of tuple
). This is sort of a compromise between a full-fledged object and a tuple, and it can be useful in those cases where you don't need the full power of a custom object, but you want your code to be more readable by avoiding weird indexing. Another use case is when there is a chance that items in the tuple need to change their position after refactoring, forcing the coder to refactor also all the logic involved, which can be very tricky. As usual, an example is better than a thousand words (or was it a picture?). Say we are handling data about the left and right eye of a patient. We save one value for the left eye (position 0) and one for the right eye (position 1) in a regular tuple. Here's how that might be:
>>> vision = (9.5, 8.8) >>> vision (9.5, 8.8) >>> vision[0] # left eye (implicit positional reference) 9.5 >>> vision[1] # right eye (implicit positional reference) 8.8
Now let's pretend we handle vision
object all the time, and at some point the designer decides to enhance them by adding information for the combined vision, so that a vision
object stores data in this format: (left eye, combined, right eye).
Do you see the trouble we're in now? We may have a lot of code that depends on vision[0]
being the left eye information (which still is) and vision[1]
being the right eye information (which is no longer the case). We have to refactor our code wherever we handle these objects, changing vision[1]
to vision[2]
, and it can be painful. We could have probably approached this a bit better from the beginning, by using a namedtuple
. Let me show you what I mean:
>>> from collections import namedtuple >>> Vision = namedtuple('Vision', ['left', 'right']) >>> vision = Vision(9.5, 8.8) >>> vision[0] 9.5 >>> vision.left # same as vision[0], but explicit 9.5 >>> vision.right # same as vision[1], but explicit 8.8
If within our code we refer to left and right eye using vision.left
and vision.right
, all we need to do to fix the new design issue is to change our factory and the way we create instances. The rest of the code won't need to change.
>>> Vision = namedtuple('Vision', ['left', 'combined', 'right']) >>> vision = Vision(9.5, 9.2, 8.8) >>> vision.left # still perfect 9.5 >>> vision.right # still perfect (though now is vision[2]) 8.8 >>> vision.combined # the new vision[1] 9.2
You can see how convenient it is to refer to those values by name rather than by position. After all, a wise man once wrote "Explicit is better than implicit" (can you recall where? Think zen if you don't...). This example may be a little extreme, of course it's not likely that our code designer will go for a change like this, but you'd be amazed to see how frequently issues similar to this one happen in a professional environment, and how painful it is to refactor them.
The defaultdict
data type is one of my favorites. It allows you to avoid checking if a key is in a dictionary by simply inserting it for you on your first access attempt, with a default value whose type you pass on creation. In some cases, this tool can be very handy and shorten your code a little. Let's see a quick example: say we are updating the value of age
, by adding one year. If age
is not there, we assume it was 0
and we update it to 1
.
>>> d = {} >>> d['age'] = d.get('age', 0) + 1 # age not there, we get 0 + 1 >>> d {'age': 1} >>> d = {'age': 39} >>> d['age'] = d.get('age', 0) + 1 # age is there, we get 40 >>> d {'age': 40}
Now let's see how it would work with a defaultdict
data type. The second line is actually the short version of a 4-lines long if
clause that we would have to write if dictionaries didn't have the get
method. We'll see all about if
clauses in Chapter 3, Iterating and Making Decisions.
>>> from collections import defaultdict >>> dd = defaultdict(int) # int is the default type (0 the value) >>> dd['age'] += 1 # short for dd['age'] = dd['age'] + 1 >>> dd defaultdict(<class 'int'>, {'age': 1}) # 1, as expected >>> dd['age'] = 39 >>> dd['age'] += 1 >>> dd defaultdict(<class 'int'>, {'age': 40}) # 40, as expected
Notice how we just need to instruct the defaultdict
factory that we want an int
number to be used in case the key is missing (we'll get 0
, which is the default for the int
type). Also, notice that even though in this example there is no gain on the number of lines, there is definitely a gain in readability, which is very important. You can also use a different technique to instantiate a defaultdict
data type, which involves creating a factory object. For digging deeper, please refer to the official documentation.
The ChainMap
is an extremely nice data type which was introduced in Python 3.3. It behaves like a normal dictionary but according to the Python documentation: is provided for quickly linking a number of mappings so they can be treated as a single unit. This is usually much faster than creating one dictionary and running multiple update calls on it. ChainMap
can be used to simulate nested scopes and is useful in templating. The underlying mappings are stored in a list. That list is public and can be accessed or updated using the maps attribute. Lookups search the underlying mappings successively until a key is found. In contrast, writes, updates, and deletions only operate on the first mapping.
A very common use case is providing defaults, so let's see an example:
>>> from collections import ChainMap >>> default_connection = {'host': 'localhost', 'port': 4567} >>> connection = {'port': 5678} >>> conn = ChainMap(connection, default_connection) # map creation >>> conn['port'] # port is found in the first dictionary 5678 >>> conn['host'] # host is fetched from the second dictionary 'localhost' >>> conn.maps # we can see the mapping objects [{'port': 5678}, {'host': 'localhost', 'port': 4567}] >>> conn['host'] = 'packtpub.com' # let's add host >>> conn.maps [{'host': 'packtpub.com', 'port': 5678}, {'host': 'localhost', 'port': 4567}] >>> del conn['port'] # let's remove the port information >>> conn.maps [{'host': 'packtpub.com'}, {'host': 'localhost', 'port': 4567}] >>> conn['port'] # now port is fetched from the second dictionary 4567 >>> dict(conn) # easy to merge and convert to regular dictionary {'host': 'packtpub.com', 'port': 4567}
I just love how Python makes your life easy. You work on a ChainMap
object, configure the first mapping as you want, and when you need a complete dictionary with all the defaults as well as the customized items, you just feed the ChainMap
object to a dict
constructor. If you have never coded in other languages, such as Java or C++, you probably won't be able to fully appreciate how precious this is, how Python makes your life so much easier. I do, I feel claustrophobic every time I have to code in some other language.
Named tuples
A namedtuple
is a tuple
-like object that has fields accessible by attribute lookup as well as being indexable and iterable (it's actually a subclass of tuple
). This is sort of a compromise between a full-fledged object and a tuple, and it can be useful in those cases where you don't need the full power of a custom object, but you want your code to be more readable by avoiding weird indexing. Another use case is when there is a chance that items in the tuple need to change their position after refactoring, forcing the coder to refactor also all the logic involved, which can be very tricky. As usual, an example is better than a thousand words (or was it a picture?). Say we are handling data about the left and right eye of a patient. We save one value for the left eye (position 0) and one for the right eye (position 1) in a regular tuple. Here's how that might be:
>>> vision = (9.5, 8.8) >>> vision (9.5, 8.8) >>> vision[0] # left eye (implicit positional reference) 9.5 >>> vision[1] # right eye (implicit positional reference) 8.8
Now let's pretend we handle vision
object all the time, and at some point the designer decides to enhance them by adding information for the combined vision, so that a vision
object stores data in this format: (left eye, combined, right eye).
Do you see the trouble we're in now? We may have a lot of code that depends on vision[0]
being the left eye information (which still is) and vision[1]
being the right eye information (which is no longer the case). We have to refactor our code wherever we handle these objects, changing vision[1]
to vision[2]
, and it can be painful. We could have probably approached this a bit better from the beginning, by using a namedtuple
. Let me show you what I mean:
>>> from collections import namedtuple >>> Vision = namedtuple('Vision', ['left', 'right']) >>> vision = Vision(9.5, 8.8) >>> vision[0] 9.5 >>> vision.left # same as vision[0], but explicit 9.5 >>> vision.right # same as vision[1], but explicit 8.8
If within our code we refer to left and right eye using vision.left
and vision.right
, all we need to do to fix the new design issue is to change our factory and the way we create instances. The rest of the code won't need to change.
>>> Vision = namedtuple('Vision', ['left', 'combined', 'right']) >>> vision = Vision(9.5, 9.2, 8.8) >>> vision.left # still perfect 9.5 >>> vision.right # still perfect (though now is vision[2]) 8.8 >>> vision.combined # the new vision[1] 9.2
You can see how convenient it is to refer to those values by name rather than by position. After all, a wise man once wrote "Explicit is better than implicit" (can you recall where? Think zen if you don't...). This example may be a little extreme, of course it's not likely that our code designer will go for a change like this, but you'd be amazed to see how frequently issues similar to this one happen in a professional environment, and how painful it is to refactor them.
The defaultdict
data type is one of my favorites. It allows you to avoid checking if a key is in a dictionary by simply inserting it for you on your first access attempt, with a default value whose type you pass on creation. In some cases, this tool can be very handy and shorten your code a little. Let's see a quick example: say we are updating the value of age
, by adding one year. If age
is not there, we assume it was 0
and we update it to 1
.
>>> d = {} >>> d['age'] = d.get('age', 0) + 1 # age not there, we get 0 + 1 >>> d {'age': 1} >>> d = {'age': 39} >>> d['age'] = d.get('age', 0) + 1 # age is there, we get 40 >>> d {'age': 40}
Now let's see how it would work with a defaultdict
data type. The second line is actually the short version of a 4-lines long if
clause that we would have to write if dictionaries didn't have the get
method. We'll see all about if
clauses in Chapter 3, Iterating and Making Decisions.
>>> from collections import defaultdict >>> dd = defaultdict(int) # int is the default type (0 the value) >>> dd['age'] += 1 # short for dd['age'] = dd['age'] + 1 >>> dd defaultdict(<class 'int'>, {'age': 1}) # 1, as expected >>> dd['age'] = 39 >>> dd['age'] += 1 >>> dd defaultdict(<class 'int'>, {'age': 40}) # 40, as expected
Notice how we just need to instruct the defaultdict
factory that we want an int
number to be used in case the key is missing (we'll get 0
, which is the default for the int
type). Also, notice that even though in this example there is no gain on the number of lines, there is definitely a gain in readability, which is very important. You can also use a different technique to instantiate a defaultdict
data type, which involves creating a factory object. For digging deeper, please refer to the official documentation.
The ChainMap
is an extremely nice data type which was introduced in Python 3.3. It behaves like a normal dictionary but according to the Python documentation: is provided for quickly linking a number of mappings so they can be treated as a single unit. This is usually much faster than creating one dictionary and running multiple update calls on it. ChainMap
can be used to simulate nested scopes and is useful in templating. The underlying mappings are stored in a list. That list is public and can be accessed or updated using the maps attribute. Lookups search the underlying mappings successively until a key is found. In contrast, writes, updates, and deletions only operate on the first mapping.
A very common use case is providing defaults, so let's see an example:
>>> from collections import ChainMap >>> default_connection = {'host': 'localhost', 'port': 4567} >>> connection = {'port': 5678} >>> conn = ChainMap(connection, default_connection) # map creation >>> conn['port'] # port is found in the first dictionary 5678 >>> conn['host'] # host is fetched from the second dictionary 'localhost' >>> conn.maps # we can see the mapping objects [{'port': 5678}, {'host': 'localhost', 'port': 4567}] >>> conn['host'] = 'packtpub.com' # let's add host >>> conn.maps [{'host': 'packtpub.com', 'port': 5678}, {'host': 'localhost', 'port': 4567}] >>> del conn['port'] # let's remove the port information >>> conn.maps [{'host': 'packtpub.com'}, {'host': 'localhost', 'port': 4567}] >>> conn['port'] # now port is fetched from the second dictionary 4567 >>> dict(conn) # easy to merge and convert to regular dictionary {'host': 'packtpub.com', 'port': 4567}
I just love how Python makes your life easy. You work on a ChainMap
object, configure the first mapping as you want, and when you need a complete dictionary with all the defaults as well as the customized items, you just feed the ChainMap
object to a dict
constructor. If you have never coded in other languages, such as Java or C++, you probably won't be able to fully appreciate how precious this is, how Python makes your life so much easier. I do, I feel claustrophobic every time I have to code in some other language.
Defaultdict
The defaultdict
data type is one of my favorites. It allows you to avoid checking if a key is in a dictionary by simply inserting it for you on your first access attempt, with a default value whose type you pass on creation. In some cases, this tool can be very handy and shorten your code a little. Let's see a quick example: say we are updating the value of age
, by adding one year. If age
is not there, we assume it was 0
and we update it to 1
.
>>> d = {} >>> d['age'] = d.get('age', 0) + 1 # age not there, we get 0 + 1 >>> d {'age': 1} >>> d = {'age': 39} >>> d['age'] = d.get('age', 0) + 1 # age is there, we get 40 >>> d {'age': 40}
Now let's see how it would work with a defaultdict
data type. The second line is actually the short version of a 4-lines long if
clause that we would have to write if dictionaries didn't have the get
method. We'll see all about if
clauses in Chapter 3, Iterating and Making Decisions.
>>> from collections import defaultdict >>> dd = defaultdict(int) # int is the default type (0 the value) >>> dd['age'] += 1 # short for dd['age'] = dd['age'] + 1 >>> dd defaultdict(<class 'int'>, {'age': 1}) # 1, as expected >>> dd['age'] = 39 >>> dd['age'] += 1 >>> dd defaultdict(<class 'int'>, {'age': 40}) # 40, as expected
Notice how we just need to instruct the defaultdict
factory that we want an int
number to be used in case the key is missing (we'll get 0
, which is the default for the int
type). Also, notice that even though in this example there is no gain on the number of lines, there is definitely a gain in readability, which is very important. You can also use a different technique to instantiate a defaultdict
data type, which involves creating a factory object. For digging deeper, please refer to the official documentation.
The ChainMap
is an extremely nice data type which was introduced in Python 3.3. It behaves like a normal dictionary but according to the Python documentation: is provided for quickly linking a number of mappings so they can be treated as a single unit. This is usually much faster than creating one dictionary and running multiple update calls on it. ChainMap
can be used to simulate nested scopes and is useful in templating. The underlying mappings are stored in a list. That list is public and can be accessed or updated using the maps attribute. Lookups search the underlying mappings successively until a key is found. In contrast, writes, updates, and deletions only operate on the first mapping.
A very common use case is providing defaults, so let's see an example:
>>> from collections import ChainMap >>> default_connection = {'host': 'localhost', 'port': 4567} >>> connection = {'port': 5678} >>> conn = ChainMap(connection, default_connection) # map creation >>> conn['port'] # port is found in the first dictionary 5678 >>> conn['host'] # host is fetched from the second dictionary 'localhost' >>> conn.maps # we can see the mapping objects [{'port': 5678}, {'host': 'localhost', 'port': 4567}] >>> conn['host'] = 'packtpub.com' # let's add host >>> conn.maps [{'host': 'packtpub.com', 'port': 5678}, {'host': 'localhost', 'port': 4567}] >>> del conn['port'] # let's remove the port information >>> conn.maps [{'host': 'packtpub.com'}, {'host': 'localhost', 'port': 4567}] >>> conn['port'] # now port is fetched from the second dictionary 4567 >>> dict(conn) # easy to merge and convert to regular dictionary {'host': 'packtpub.com', 'port': 4567}
I just love how Python makes your life easy. You work on a ChainMap
object, configure the first mapping as you want, and when you need a complete dictionary with all the defaults as well as the customized items, you just feed the ChainMap
object to a dict
constructor. If you have never coded in other languages, such as Java or C++, you probably won't be able to fully appreciate how precious this is, how Python makes your life so much easier. I do, I feel claustrophobic every time I have to code in some other language.
ChainMap
The ChainMap
is an extremely nice data type which was introduced in Python 3.3. It behaves like a normal dictionary but according to the Python documentation: is provided for quickly linking a number of mappings so they can be treated as a single unit. This is usually much faster than creating one dictionary and running multiple update calls on it. ChainMap
can be used to simulate nested scopes and is useful in templating. The underlying mappings are stored in a list. That list is public and can be accessed or updated using the maps attribute. Lookups search the underlying mappings successively until a key is found. In contrast, writes, updates, and deletions only operate on the first mapping.
A very common use case is providing defaults, so let's see an example:
>>> from collections import ChainMap >>> default_connection = {'host': 'localhost', 'port': 4567} >>> connection = {'port': 5678} >>> conn = ChainMap(connection, default_connection) # map creation >>> conn['port'] # port is found in the first dictionary 5678 >>> conn['host'] # host is fetched from the second dictionary 'localhost' >>> conn.maps # we can see the mapping objects [{'port': 5678}, {'host': 'localhost', 'port': 4567}] >>> conn['host'] = 'packtpub.com' # let's add host >>> conn.maps [{'host': 'packtpub.com', 'port': 5678}, {'host': 'localhost', 'port': 4567}] >>> del conn['port'] # let's remove the port information >>> conn.maps [{'host': 'packtpub.com'}, {'host': 'localhost', 'port': 4567}] >>> conn['port'] # now port is fetched from the second dictionary 4567 >>> dict(conn) # easy to merge and convert to regular dictionary {'host': 'packtpub.com', 'port': 4567}
I just love how Python makes your life easy. You work on a ChainMap
object, configure the first mapping as you want, and when you need a complete dictionary with all the defaults as well as the customized items, you just feed the ChainMap
object to a dict
constructor. If you have never coded in other languages, such as Java or C++, you probably won't be able to fully appreciate how precious this is, how Python makes your life so much easier. I do, I feel claustrophobic every time I have to code in some other language.
That's it. Now you have seen a very good portion of the data structures that you will use in Python. I encourage you to take a dive into the Python documentation and experiment further with each and every data type we've seen in this chapter. It's worth it, believe me. Everything you'll write will be about handling data, so make sure your knowledge about it is rock solid.
Before we leap into the next chapter, I'd like to make some final considerations about different aspects that to my mind are important and not to be neglected.
When we discussed objects at the beginning of this chapter, we saw that when we assigned a name to an object, Python creates the object, sets its value, and then points the name to it. We can assign different names to the same value and we expect different objects to be created, like this:
>>> a = 1000000 >>> b = 1000000 >>> id(a) == id(b) False
In the preceding example, a
and b
are assigned to two int
objects, which have the same value but they are not the same object, as you can see, their id
is not the same. So let's do it again:
>>> a = 5 >>> b = 5 >>> id(a) == id(b) True
Oh oh! Is Python broken? Why are the two objects the same now? We didn't do a = b = 5
, we set them up separately. Well, the answer is performances. Python caches short strings and small numbers, to avoid having many copies of them clogging up the system memory. Everything is handled properly under the hood so you don't need to worry a bit, but make sure that you remember this behavior should your code ever need to fiddle with IDs.
As we've seen, Python provides you with several built-in data types and sometimes, if you're not that experienced, choosing the one that serves you best can be tricky, especially when it comes to collections. For example, say you have many dictionaries to store, each of which represents a customer. Within each customer dictionary there's an 'id': 'code'
unique identification code. In what kind of collection would you place them? Well, unless I know more about these customers, it's very hard to answer. What kind of access will I need? What sort of operations will I have to perform on each of them, and how many times? Will the collection change over time? Will I need to modify the customer dictionaries in any way? What is going to be the most frequent operation I will have to perform on the collection?
If you can answer the preceding questions, then you will know what to choose. If the collection never shrinks or grows (in other words, it won't need to add/delete any customer object after creation) or shuffles, then tuples are a possible choice. Otherwise lists are a good candidate. Every customer dictionary has a unique identifier though, so even a dictionary could work. Let me draft these options for you:
# example customer objects customer1 = {'id': 'abc123', 'full_name': 'Master Yoda'} customer2 = {'id': 'def456', 'full_name': 'Obi-Wan Kenobi'} customer3 = {'id': 'ghi789', 'full_name': 'Anakin Skywalker'} # collect them in a tuple customers = (customer1, customer2, customer3) # or collect them in a list customers = [customer1, customer2, customer3] # or maybe within a dictionary, they have a unique id after all customers = { 'abc123': customer1, 'def456': customer2, 'ghi789': customer3, }
Some customers we have there, right? I probably wouldn't go with the tuple option, unless I wanted to highlight that the collection is not going to change. I'd say usually a list is better, it allows for more flexibility.
Another factor to keep in mind is that tuples and lists are ordered collections, while if you use a dictionary or a set you lose the ordering, so you need to know if ordering is important in your application.
What about performances? For example in a list, operations such as insertion and membership can take O(n), while they are O(1) for a dictionary. It's not always possible to use dictionaries though, if we don't have the guarantee that we can uniquely identify each item of the collection by means of one of its properties, and that the property in question is hashable (so it can be a key in dict
).
Note
If you're wondering what O(n) and O(1) mean, please Google "big O notation" and get a gist of it from anywhere. In this context, let's just say that if performing an operation Op on a data structure takes O(f(n)), it would mean that Op takes at most a time to complete, where c is some positive constant, n is the size of the input, and f is some function. So, think of O(...) as an upper bound for the running time of an operation (it can be used also to size other measurable quantities, of course).
Another way of understanding if you have chosen the right data structure is by looking at the code you have to write in order to manipulate it. If everything comes easily and flows naturally, then you probably have chosen correctly, but if you find yourself thinking your code is getting unnecessarily complicated, then you probably should try and decide whether you need to reconsider your choices. It's quite hard to give advice without a practical case though, so when you choose a data structure for your data, try to keep ease of use and performance in mind and give precedence to what matters most in the context you are.
At the beginning of this chapter, we saw slicing applied on strings. Slicing in general applies to a sequence, so tuples, lists, strings, etc. With lists, slicing can also be used for assignment. I've almost never seen this used in professional code, but still, you know you can. Could you slice dictionaries or sets? I hear you scream "Of course not! They are not ordered!". Excellent, I see we're on the same page here, so let's talk about indexing.
There is one characteristic about Python indexing I haven't mentioned before. I'll show you by example. How do you address the last element of a collection? Let's see:
>>> a = list(range(10)) # `a` has 10 elements. Last one is 9. >>> a [0, 1, 2, 3, 4, 5, 6, 7, 8, 9] >>> len(a) # its length is 10 elements 10 >>> a[len(a) - 1] # position of last one is len(a) - 1 9 >>> a[-1] # but we don't need len(a)! Python rocks! 9 >>> a[-2] # equivalent to len(a) - 2 8 >>> a[-3] # equivalent to len(a) - 3 7
If the list a
has 10 elements, because of the 0-index positioning system of Python, the first one is at position 0 and the last one is at position 9. In the preceding example, the elements are conveniently placed in a position equal to their value: 0 is at position 0, 1 at position 1, and so on.
So, in order to fetch the last element, we need to know the length of the whole list (or tuple, or string, and so on) and then subtract 1. Hence: len(a) – 1
. This is so common an operation that Python provides you with a way to retrieve elements using negative indexing. This proves very useful when you do some serious data manipulation. Here's a nice diagram about how indexing works on the string "HelloThere"
:

Trying to address indexes greater than 9 or smaller than -10 will raise an IndexError
, as expected.
You may have noticed that, in order to keep the example as short as possible, I have called many objects using simple letters, like a
, b
, c
, d
, and so on. This is perfectly ok when you debug on the console or when you show that a + b == 7
, but it's bad practice when it comes to professional coding (or any type of coding, for all that matter). I hope you will indulge me if I sometimes do it, the reason is to present the code in a more compact way.
In a real environment though, when you choose names for your data, you should choose them carefully and they should reflect what the data is about. So, if you have a collection of Customer
objects, customers
is a perfectly good name for it. Would customers_list
, customers_tuple
, or customers_collection
work as well? Think about it for a second. Is it good to tie the name of the collection to the data type? I don't think so, at least in most cases. So I'd say if you have an excellent reason to do so go ahead, otherwise don't. The reason is, once that customers_tuple
starts being used in different places of your code, and you realize you actually want to use a list instead of a tuple, you're up for some fun refactoring (also known as wasted time). Names for data should be nouns, and names for functions should be verbs. Names should be as expressive as possible. Python is actually a very good example when it comes to names. Most of the time you can just guess what a function is called if you know what it does. Crazy, huh?
Chapter 2, Meaningful Names of Clean Code, Robert C. Martin, Prentice Hall is entirely dedicated to names. It's an amazing book that helped me improve my coding style in many different ways, a must read if you want to take your coding to the next level.
Small values caching
When we discussed objects at the beginning of this chapter, we saw that when we assigned a name to an object, Python creates the object, sets its value, and then points the name to it. We can assign different names to the same value and we expect different objects to be created, like this:
>>> a = 1000000 >>> b = 1000000 >>> id(a) == id(b) False
In the preceding example, a
and b
are assigned to two int
objects, which have the same value but they are not the same object, as you can see, their id
is not the same. So let's do it again:
>>> a = 5 >>> b = 5 >>> id(a) == id(b) True
Oh oh! Is Python broken? Why are the two objects the same now? We didn't do a = b = 5
, we set them up separately. Well, the answer is performances. Python caches short strings and small numbers, to avoid having many copies of them clogging up the system memory. Everything is handled properly under the hood so you don't need to worry a bit, but make sure that you remember this behavior should your code ever need to fiddle with IDs.
As we've seen, Python provides you with several built-in data types and sometimes, if you're not that experienced, choosing the one that serves you best can be tricky, especially when it comes to collections. For example, say you have many dictionaries to store, each of which represents a customer. Within each customer dictionary there's an 'id': 'code'
unique identification code. In what kind of collection would you place them? Well, unless I know more about these customers, it's very hard to answer. What kind of access will I need? What sort of operations will I have to perform on each of them, and how many times? Will the collection change over time? Will I need to modify the customer dictionaries in any way? What is going to be the most frequent operation I will have to perform on the collection?
If you can answer the preceding questions, then you will know what to choose. If the collection never shrinks or grows (in other words, it won't need to add/delete any customer object after creation) or shuffles, then tuples are a possible choice. Otherwise lists are a good candidate. Every customer dictionary has a unique identifier though, so even a dictionary could work. Let me draft these options for you:
# example customer objects customer1 = {'id': 'abc123', 'full_name': 'Master Yoda'} customer2 = {'id': 'def456', 'full_name': 'Obi-Wan Kenobi'} customer3 = {'id': 'ghi789', 'full_name': 'Anakin Skywalker'} # collect them in a tuple customers = (customer1, customer2, customer3) # or collect them in a list customers = [customer1, customer2, customer3] # or maybe within a dictionary, they have a unique id after all customers = { 'abc123': customer1, 'def456': customer2, 'ghi789': customer3, }
Some customers we have there, right? I probably wouldn't go with the tuple option, unless I wanted to highlight that the collection is not going to change. I'd say usually a list is better, it allows for more flexibility.
Another factor to keep in mind is that tuples and lists are ordered collections, while if you use a dictionary or a set you lose the ordering, so you need to know if ordering is important in your application.
What about performances? For example in a list, operations such as insertion and membership can take O(n), while they are O(1) for a dictionary. It's not always possible to use dictionaries though, if we don't have the guarantee that we can uniquely identify each item of the collection by means of one of its properties, and that the property in question is hashable (so it can be a key in dict
).
Note
If you're wondering what O(n) and O(1) mean, please Google "big O notation" and get a gist of it from anywhere. In this context, let's just say that if performing an operation Op on a data structure takes O(f(n)), it would mean that Op takes at most a time to complete, where c is some positive constant, n is the size of the input, and f is some function. So, think of O(...) as an upper bound for the running time of an operation (it can be used also to size other measurable quantities, of course).
Another way of understanding if you have chosen the right data structure is by looking at the code you have to write in order to manipulate it. If everything comes easily and flows naturally, then you probably have chosen correctly, but if you find yourself thinking your code is getting unnecessarily complicated, then you probably should try and decide whether you need to reconsider your choices. It's quite hard to give advice without a practical case though, so when you choose a data structure for your data, try to keep ease of use and performance in mind and give precedence to what matters most in the context you are.
At the beginning of this chapter, we saw slicing applied on strings. Slicing in general applies to a sequence, so tuples, lists, strings, etc. With lists, slicing can also be used for assignment. I've almost never seen this used in professional code, but still, you know you can. Could you slice dictionaries or sets? I hear you scream "Of course not! They are not ordered!". Excellent, I see we're on the same page here, so let's talk about indexing.
There is one characteristic about Python indexing I haven't mentioned before. I'll show you by example. How do you address the last element of a collection? Let's see:
>>> a = list(range(10)) # `a` has 10 elements. Last one is 9. >>> a [0, 1, 2, 3, 4, 5, 6, 7, 8, 9] >>> len(a) # its length is 10 elements 10 >>> a[len(a) - 1] # position of last one is len(a) - 1 9 >>> a[-1] # but we don't need len(a)! Python rocks! 9 >>> a[-2] # equivalent to len(a) - 2 8 >>> a[-3] # equivalent to len(a) - 3 7
If the list a
has 10 elements, because of the 0-index positioning system of Python, the first one is at position 0 and the last one is at position 9. In the preceding example, the elements are conveniently placed in a position equal to their value: 0 is at position 0, 1 at position 1, and so on.
So, in order to fetch the last element, we need to know the length of the whole list (or tuple, or string, and so on) and then subtract 1. Hence: len(a) – 1
. This is so common an operation that Python provides you with a way to retrieve elements using negative indexing. This proves very useful when you do some serious data manipulation. Here's a nice diagram about how indexing works on the string "HelloThere"
:

Trying to address indexes greater than 9 or smaller than -10 will raise an IndexError
, as expected.
You may have noticed that, in order to keep the example as short as possible, I have called many objects using simple letters, like a
, b
, c
, d
, and so on. This is perfectly ok when you debug on the console or when you show that a + b == 7
, but it's bad practice when it comes to professional coding (or any type of coding, for all that matter). I hope you will indulge me if I sometimes do it, the reason is to present the code in a more compact way.
In a real environment though, when you choose names for your data, you should choose them carefully and they should reflect what the data is about. So, if you have a collection of Customer
objects, customers
is a perfectly good name for it. Would customers_list
, customers_tuple
, or customers_collection
work as well? Think about it for a second. Is it good to tie the name of the collection to the data type? I don't think so, at least in most cases. So I'd say if you have an excellent reason to do so go ahead, otherwise don't. The reason is, once that customers_tuple
starts being used in different places of your code, and you realize you actually want to use a list instead of a tuple, you're up for some fun refactoring (also known as wasted time). Names for data should be nouns, and names for functions should be verbs. Names should be as expressive as possible. Python is actually a very good example when it comes to names. Most of the time you can just guess what a function is called if you know what it does. Crazy, huh?
Chapter 2, Meaningful Names of Clean Code, Robert C. Martin, Prentice Hall is entirely dedicated to names. It's an amazing book that helped me improve my coding style in many different ways, a must read if you want to take your coding to the next level.
How to choose data structures
As we've seen, Python provides you with several built-in data types and sometimes, if you're not that experienced, choosing the one that serves you best can be tricky, especially when it comes to collections. For example, say you have many dictionaries to store, each of which represents a customer. Within each customer dictionary there's an 'id': 'code'
unique identification code. In what kind of collection would you place them? Well, unless I know more about these customers, it's very hard to answer. What kind of access will I need? What sort of operations will I have to perform on each of them, and how many times? Will the collection change over time? Will I need to modify the customer dictionaries in any way? What is going to be the most frequent operation I will have to perform on the collection?
If you can answer the preceding questions, then you will know what to choose. If the collection never shrinks or grows (in other words, it won't need to add/delete any customer object after creation) or shuffles, then tuples are a possible choice. Otherwise lists are a good candidate. Every customer dictionary has a unique identifier though, so even a dictionary could work. Let me draft these options for you:
# example customer objects customer1 = {'id': 'abc123', 'full_name': 'Master Yoda'} customer2 = {'id': 'def456', 'full_name': 'Obi-Wan Kenobi'} customer3 = {'id': 'ghi789', 'full_name': 'Anakin Skywalker'} # collect them in a tuple customers = (customer1, customer2, customer3) # or collect them in a list customers = [customer1, customer2, customer3] # or maybe within a dictionary, they have a unique id after all customers = { 'abc123': customer1, 'def456': customer2, 'ghi789': customer3, }
Some customers we have there, right? I probably wouldn't go with the tuple option, unless I wanted to highlight that the collection is not going to change. I'd say usually a list is better, it allows for more flexibility.
Another factor to keep in mind is that tuples and lists are ordered collections, while if you use a dictionary or a set you lose the ordering, so you need to know if ordering is important in your application.
What about performances? For example in a list, operations such as insertion and membership can take O(n), while they are O(1) for a dictionary. It's not always possible to use dictionaries though, if we don't have the guarantee that we can uniquely identify each item of the collection by means of one of its properties, and that the property in question is hashable (so it can be a key in dict
).
Note
If you're wondering what O(n) and O(1) mean, please Google "big O notation" and get a gist of it from anywhere. In this context, let's just say that if performing an operation Op on a data structure takes O(f(n)), it would mean that Op takes at most a time to complete, where c is some positive constant, n is the size of the input, and f is some function. So, think of O(...) as an upper bound for the running time of an operation (it can be used also to size other measurable quantities, of course).
Another way of understanding if you have chosen the right data structure is by looking at the code you have to write in order to manipulate it. If everything comes easily and flows naturally, then you probably have chosen correctly, but if you find yourself thinking your code is getting unnecessarily complicated, then you probably should try and decide whether you need to reconsider your choices. It's quite hard to give advice without a practical case though, so when you choose a data structure for your data, try to keep ease of use and performance in mind and give precedence to what matters most in the context you are.
At the beginning of this chapter, we saw slicing applied on strings. Slicing in general applies to a sequence, so tuples, lists, strings, etc. With lists, slicing can also be used for assignment. I've almost never seen this used in professional code, but still, you know you can. Could you slice dictionaries or sets? I hear you scream "Of course not! They are not ordered!". Excellent, I see we're on the same page here, so let's talk about indexing.
There is one characteristic about Python indexing I haven't mentioned before. I'll show you by example. How do you address the last element of a collection? Let's see:
>>> a = list(range(10)) # `a` has 10 elements. Last one is 9. >>> a [0, 1, 2, 3, 4, 5, 6, 7, 8, 9] >>> len(a) # its length is 10 elements 10 >>> a[len(a) - 1] # position of last one is len(a) - 1 9 >>> a[-1] # but we don't need len(a)! Python rocks! 9 >>> a[-2] # equivalent to len(a) - 2 8 >>> a[-3] # equivalent to len(a) - 3 7
If the list a
has 10 elements, because of the 0-index positioning system of Python, the first one is at position 0 and the last one is at position 9. In the preceding example, the elements are conveniently placed in a position equal to their value: 0 is at position 0, 1 at position 1, and so on.
So, in order to fetch the last element, we need to know the length of the whole list (or tuple, or string, and so on) and then subtract 1. Hence: len(a) – 1
. This is so common an operation that Python provides you with a way to retrieve elements using negative indexing. This proves very useful when you do some serious data manipulation. Here's a nice diagram about how indexing works on the string "HelloThere"
:

Trying to address indexes greater than 9 or smaller than -10 will raise an IndexError
, as expected.
You may have noticed that, in order to keep the example as short as possible, I have called many objects using simple letters, like a
, b
, c
, d
, and so on. This is perfectly ok when you debug on the console or when you show that a + b == 7
, but it's bad practice when it comes to professional coding (or any type of coding, for all that matter). I hope you will indulge me if I sometimes do it, the reason is to present the code in a more compact way.
In a real environment though, when you choose names for your data, you should choose them carefully and they should reflect what the data is about. So, if you have a collection of Customer
objects, customers
is a perfectly good name for it. Would customers_list
, customers_tuple
, or customers_collection
work as well? Think about it for a second. Is it good to tie the name of the collection to the data type? I don't think so, at least in most cases. So I'd say if you have an excellent reason to do so go ahead, otherwise don't. The reason is, once that customers_tuple
starts being used in different places of your code, and you realize you actually want to use a list instead of a tuple, you're up for some fun refactoring (also known as wasted time). Names for data should be nouns, and names for functions should be verbs. Names should be as expressive as possible. Python is actually a very good example when it comes to names. Most of the time you can just guess what a function is called if you know what it does. Crazy, huh?
Chapter 2, Meaningful Names of Clean Code, Robert C. Martin, Prentice Hall is entirely dedicated to names. It's an amazing book that helped me improve my coding style in many different ways, a must read if you want to take your coding to the next level.
About indexing and slicing
At the beginning of this chapter, we saw slicing applied on strings. Slicing in general applies to a sequence, so tuples, lists, strings, etc. With lists, slicing can also be used for assignment. I've almost never seen this used in professional code, but still, you know you can. Could you slice dictionaries or sets? I hear you scream "Of course not! They are not ordered!". Excellent, I see we're on the same page here, so let's talk about indexing.
There is one characteristic about Python indexing I haven't mentioned before. I'll show you by example. How do you address the last element of a collection? Let's see:
>>> a = list(range(10)) # `a` has 10 elements. Last one is 9. >>> a [0, 1, 2, 3, 4, 5, 6, 7, 8, 9] >>> len(a) # its length is 10 elements 10 >>> a[len(a) - 1] # position of last one is len(a) - 1 9 >>> a[-1] # but we don't need len(a)! Python rocks! 9 >>> a[-2] # equivalent to len(a) - 2 8 >>> a[-3] # equivalent to len(a) - 3 7
If the list a
has 10 elements, because of the 0-index positioning system of Python, the first one is at position 0 and the last one is at position 9. In the preceding example, the elements are conveniently placed in a position equal to their value: 0 is at position 0, 1 at position 1, and so on.
So, in order to fetch the last element, we need to know the length of the whole list (or tuple, or string, and so on) and then subtract 1. Hence: len(a) – 1
. This is so common an operation that Python provides you with a way to retrieve elements using negative indexing. This proves very useful when you do some serious data manipulation. Here's a nice diagram about how indexing works on the string "HelloThere"
:

Trying to address indexes greater than 9 or smaller than -10 will raise an IndexError
, as expected.
You may have noticed that, in order to keep the example as short as possible, I have called many objects using simple letters, like a
, b
, c
, d
, and so on. This is perfectly ok when you debug on the console or when you show that a + b == 7
, but it's bad practice when it comes to professional coding (or any type of coding, for all that matter). I hope you will indulge me if I sometimes do it, the reason is to present the code in a more compact way.
In a real environment though, when you choose names for your data, you should choose them carefully and they should reflect what the data is about. So, if you have a collection of Customer
objects, customers
is a perfectly good name for it. Would customers_list
, customers_tuple
, or customers_collection
work as well? Think about it for a second. Is it good to tie the name of the collection to the data type? I don't think so, at least in most cases. So I'd say if you have an excellent reason to do so go ahead, otherwise don't. The reason is, once that customers_tuple
starts being used in different places of your code, and you realize you actually want to use a list instead of a tuple, you're up for some fun refactoring (also known as wasted time). Names for data should be nouns, and names for functions should be verbs. Names should be as expressive as possible. Python is actually a very good example when it comes to names. Most of the time you can just guess what a function is called if you know what it does. Crazy, huh?
Chapter 2, Meaningful Names of Clean Code, Robert C. Martin, Prentice Hall is entirely dedicated to names. It's an amazing book that helped me improve my coding style in many different ways, a must read if you want to take your coding to the next level.
About the names
You may have noticed that, in order to keep the example as short as possible, I have called many objects using simple letters, like a
, b
, c
, d
, and so on. This is perfectly ok when you debug on the console or when you show that a + b == 7
, but it's bad practice when it comes to professional coding (or any type of coding, for all that matter). I hope you will indulge me if I sometimes do it, the reason is to present the code in a more compact way.
In a real environment though, when you choose names for your data, you should choose them carefully and they should reflect what the data is about. So, if you have a collection of Customer
objects, customers
is a perfectly good name for it. Would customers_list
, customers_tuple
, or customers_collection
work as well? Think about it for a second. Is it good to tie the name of the collection to the data type? I don't think so, at least in most cases. So I'd say if you have an excellent reason to do so go ahead, otherwise don't. The reason is, once that customers_tuple
starts being used in different places of your code, and you realize you actually want to use a list instead of a tuple, you're up for some fun refactoring (also known as wasted time). Names for data should be nouns, and names for functions should be verbs. Names should be as expressive as possible. Python is actually a very good example when it comes to names. Most of the time you can just guess what a function is called if you know what it does. Crazy, huh?
Chapter 2, Meaningful Names of Clean Code, Robert C. Martin, Prentice Hall is entirely dedicated to names. It's an amazing book that helped me improve my coding style in many different ways, a must read if you want to take your coding to the next level.
In this chapter, we've explored the built-in data types of Python. We've seen how many they are and how much can be achieved by just using them in different combinations.
We've seen number types, sequences, sets, mappings, collections, we've seen that everything is an object, we've learned the difference between mutable and immutable, and we've also learned about slicing and indexing (and, proudly, negative indexing as well).
We've presented simple examples, but there's much more that you can learn about this subject, so stick your nose into the official documentation and explore.
Most of all, I encourage you to try out all the exercises by yourself, get your fingers using that code, build some muscle memory, and experiment, experiment, experiment. Learn what happens when you divide by zero, when you combine different number types into a single expression, when you manage strings. Play with all data types. Exercise them, break them, discover all their methods, enjoy them and learn them well, damn well.
If your foundation is not rock solid, how good can your code be? And data is the foundation for everything. Data shapes what dances around it.
The more you progress with the book, the more it's likely that you will find some discrepancies or maybe a small typo here and there in my code (or yours). You will get an error message, something will break. That's wonderful! When you code, things break all the time, you debug and fix all the time, so consider errors as useful exercises to learn something new about the language you're using, and not as failures or problems. Errors will keep coming up until your very last line of code, that's for sure, so you may as well start making your peace with them now.
The next chapter is about iterating and making decisions. We'll see how to actually put those collections in use, and take decisions based on the data we're presented with. We'll start to go a little faster now that your knowledge is building up, so make sure you're comfortable with the contents of this chapter before you move to the next one. Once more, have fun, explore, break things. It's a very good way to learn.
"Insanity: doing the same thing over and over again and expecting different results." | ||
--Albert Einstein |
In the previous chapter, we've seen Python built-in data types. Now that you're familiar with data in its many forms and shapes, it's time to start looking at how a program can use it.
According to Wikipedia:
In computer science, control flow (or alternatively, flow of control) refers to the specification of the order in which the individual statements, instructions or function calls of an imperative program are executed or evaluated.
In order to control the flow of a program, we have two main weapons: conditional programming (also known as branching) and looping. We can use them in many different combinations and variations, but in this chapter, instead of going through all possible various forms of those two constructs in a "documentation" fashion, I'd rather give you the basics and then I'll write a couple of small scripts with you. In the first one, we'll see how to create a rudimentary prime number generator, while in the second one, we'll see how to apply discounts to customers based on coupons. This way you should get a better feeling about how conditional programming and looping can be used.
Conditional programming, or branching, is something you do every day, every moment. It's about evaluating conditions: if the light is green, then I can cross, if it's raining, then I'm taking the umbrella, and if I'm late for work, then I'll call my manager.
The main tool is the if
statement, which comes in different forms and colors, but basically what it does is evaluate an expression and, based on the result, choose which part of the code to execute. As usual, let's see an example:
conditional.1.py
late = True
if late:
print('I need to call my manager!')
This is possibly the simplest example: when fed to the if
statement, late
acts as a conditional expression, which is evaluated in a Boolean context (exactly like if we were calling bool(late)
). If the result of the evaluation is True
, then we enter the body of code immediately after the if
statement. Notice that the print
instruction is indented: this means it belongs to a scope defined by the if
clause. Execution of this code yields:
$ python conditional.1.py I need to call my manager!
Since late
is True
, the print
statement was executed. Let's expand on this example:
conditional.2.py
late = False
if late:
print('I need to call my manager!') #1
else:
print('no need to call my manager...') #2
This time I set late = False
, so when I execute the code, the result is different:
$ python conditional.2.py no need to call my manager...
Depending on the result of evaluating the late
expression, we can either enter block #1
or block #2
, but not both. Block #1
is executed when late
evaluates to True
, while block #2
is executed when late
evaluates to False
. Try assigning False/True
values to the late
name, and see how the output for this code changes accordingly.
The preceding example also introduces the else
clause, which becomes very handy when we want to provide an alternative set of instructions to be executed when an expression evaluates to False
within an if
clause. The else clause is optional, as it's evident by comparing the preceding two examples.
Sometimes all you need is to do something if a condition is met (simple if
clause). Other times you need to provide an alternative, in case the condition is False
(if
/else
clause), but there are situations where you may have more than two paths to choose from, so, since calling the manager (or not calling them) is kind of a binary type of example (either you call or you don't), let's change the type of example and keep expanding. This time we decide tax percentages. If my income is less then 10k, I won't pay any taxes. If it is between 10k and 30k, I'll pay 20% taxes. If it is between 30k and 100k, I'll pay 35% taxes, and over 100k, I'll (gladly) pay 45% taxes. Let's put this all down into beautiful Python code:
taxes.py
income = 15000 if income < 10000: tax_coefficient = 0.0 #1 elif income < 30000: tax_coefficient = 0.2 #2 elif income < 100000: tax_coefficient = 0.35 #3 else: tax_coefficient = 0.45 #4 print('I will pay:', income * tax_coefficient, 'in taxes')
Executing the preceding code yields:
$ python taxes.py I will pay: 3000.0 in taxes
Let's go through the example line by line: we start by setting up the income value. In the example, my income is 15k. We enter the if
clause. Notice that this time we also introduced the elif
clause, which is a contraction for else-if
, and it's different from a bare else
clause in that it also has its own condition. So, the if
expression income < 10000
, evaluates to False
, therefore block #1
is not executed. The control passes to the next condition evaluator: elif income < 30000
. This one evaluates to True
, therefore block #2
is executed, and because of this, Python then resumes execution after the whole if
/elif
/elif
/else
clause (which we can just call if
clause from now on). There is only one instruction after the if
clause, the print
call, which tells us I will pay 3k in taxes this year (15k * 20%). Notice that the order is mandatory: if
comes first, then (optionally) as many elif
as you need, and then (optionally) an else
clause.
Interesting, right? No matter how many lines of code you may have within each block, when one of the conditions evaluates to True
, the associated block is executed and then execution resumes after the whole clause. If none of the conditions evaluates to True
(for example, income = 200000
), then the body of the else
clause would be executed (block #4
). This example expands our understanding of the behavior of the else
clause. Its block of code is executed when none of the preceding if
/elif
/.../elif
expressions has evaluated to True
.
Try to modify the value of income
until you can comfortably execute all blocks at your will (one per execution, of course). And then try the boundaries. This is crucial, whenever you have conditions expressed as equalities or inequalities (==
, !=
, <
, >
, <=
, >=
), those numbers represent boundaries. It is essential to test boundaries thoroughly. Should I allow you to drive at 18 or 17? Am I checking your age with age < 18
, or age <= 18
? You can't imagine how many times I had to fix subtle bugs that stemmed from using the wrong operator, so go ahead and experiment with the preceding code. Change some <
to <=
and set income to be one of the boundary values (10k, 30k, 100k) as well as any value in between. See how the result changes, get a good understanding of it before proceeding.
Before we move to the next topic, let's see another example that shows us how to nest if
clauses. Say your program encounters an error. If the alert system is the console, we print the error. If the alert system is an e-mail, we send it according to the severity of the error. If the alert system is anything other than console or e-mail, we don't know what to do, therefore we do nothing. Let's put this into code:
errorsalert.py
alert_system = 'console' # other value can be 'email' error_severity = 'critical' # other values: 'medium' or 'low' error_message = 'OMG! Something terrible happened!' if alert_system == 'console': print(error_message) #1 elif alert_system == 'email': if error_severity == 'critical': send_email('admin@example.com', error_message) #2 elif error_severity == 'medium': send_email('support.1@example.com', error_message) #3 else: send_email('support.2@example.com', error_message) #4
The preceding example is quite interesting, in its silliness. It shows us two nested if
clauses (outer and inner). It also shows us the outer if
clause doesn't have any else
, while the inner one does. Notice how indentation is what allows us to nest one clause within another one.
If alert_system == 'console'
, body #1
is executed, and nothing else happens. On the other hand, if alert_system == 'email'
, then we enter into another if
clause, which we called inner. In the inner if
clause, according to error_severity
, we send an e-mail to either an admin, first-level support, or second-level support (blocks #2
, #3
, and #4
). The send_email
function is not defined in this example, therefore trying to run it would give you an error. In the source code of the book, which you can download from the website, I included a trick to redirect that call to a regular print
function, just so you can experiment on the console without actually sending an e-mail. Try changing the values and see how it all works.
One last thing I would like to show you before moving on to the next subject, is the ternary operator or, in layman's terms, the short version of an if
/else
clause. When the value of a name is to be assigned according to some condition, sometimes it's easier and more readable to use the ternary operator instead of a proper if
clause. In the following example, the two code blocks do exactly the same thing:
ternary.py
order_total = 247 # GBP
# classic if/else form
if order_total > 100:
discount = 25 # GBP
else:
discount = 0 # GBP
print(order_total, discount)
# ternary operator
discount = 25 if order_total > 100 else 0
print(order_total, discount)
For simple cases like this, I find it very nice to be able to express that logic in one line instead of four. Remember, as a coder, you spend much more time reading code then writing it, so Python conciseness is invaluable.
Are you clear on how the ternary operator works? Basically is name = something if condition else something-else
. So name
is assigned something
if condition
evaluates to True
, and something-else
if condition
evaluates to False
.
Now that you know everything about controlling the path of the code, let's move on to the next subject: looping.
If you have any experience with looping in other programming languages, you will find Python's way of looping a bit different. First of all, what is looping? Looping means being able to repeat the execution of a code block more than once, according to the loop parameters we're given. There are different looping constructs, which serve different purposes, and Python has distilled all of them down to just two, which you can use to achieve everything you need. These are the for and while statements.
While it's definitely possible to do everything you need using either of them, they serve different purposes and therefore they're usually used in different contexts. We'll explore this difference thoroughly through this chapter.
The for
loop is used when looping over a sequence, like a list, tuple, or a collection of objects. Let's start with a simple example that is more like C++ style, and then let's gradually see how to achieve the same results in Python (you'll love Python's syntax).
simple.for.py
for number in [0, 1, 2, 3, 4]: print(number)
This simple snippet of code, when executed, prints all numbers from 0 to 4. The for
loop is fed the list [0, 1, 2, 3, 4]
and at each iteration, number
is given a value from the sequence (which is iterated sequentially, in order), then the body of the loop is executed (the print line). number
changes at every iteration, according to which value is coming next from the sequence. When the sequence is exhausted, the for
loop terminates, and the execution of the code resumes normally with the code after the loop.
Sometimes we need to iterate over a range of numbers, and it would be quite unpleasant to have to do so by hardcoding the list somewhere. In such cases, the range
function comes to the rescue. Let's see the equivalent of the previous snippet of code:
simple.for.py
for number in range(5): print(number)
The range function is used extensively in Python programs when it comes to creating sequences: you can call it by passing one value, which acts as stop
(counting from 0), or you can pass two values (start
and stop
), or even three (start
, stop
, and step
). Check out the following example:
>>> list(range(10)) # one value: from 0 to value (excluded) [0, 1, 2, 3, 4, 5, 6, 7, 8, 9] >>> list(range(3, 8)) # two values: from start to stop (excluded) [3, 4, 5, 6, 7] >>> list(range(-10, 10, 4)) # three values: step is added [-10, -6, -2, 2, 6]
For the moment, ignore that we need to wrap range(...)
within a list
. The range
object is a little bit special, but in this case we're just interested in understanding what are the values it will return to us. You see that the deal is the same with slicing: start
is included, stop
excluded, and optionally you can add a step
parameter, which by default is 1.
Try modifying the parameters of the range()
call in our simple.for.py
code and see what it prints, get comfortable with it.
Now we have all the tools to iterate over a sequence, so let's build on that example:
simple.for.2.py
surnames = ['Rivest', 'Shamir', 'Adleman'] for position in range(len(surnames)): print(position, surnames[position])
The preceding code adds a little bit of complexity to the game. Execution will show this result:
$ python simple.for.2.py 0 Rivest 1 Shamir 2 Adleman
Let's use the inside-out technique to break it down, ok? We start from the innermost part of what we're trying to understand, and we expand outwards. So, len(surnames)
is the length of the surnames
list: 3
. Therefore, range(len(surnames))
is actually transformed into range(3)
. This gives us the range [0, 3), which is basically a sequence (0, 1, 2)
. This means that the for
loop will run three iterations. In the first one, position
will take value 0
, while in the second one, it will take value 1
, and finally value 2
in the third and last iteration. What is (0, 1, 2)
, if not the possible indexing positions for the surnames
list? At position 0
we find 'Rivest'
, at position 1
, 'Shamir'
, and at position 2
, 'Adleman'
. If you are curious about what these three men created together, change print(position, surnames[position])
to print(surnames[position][0], end='')
add a final print()
outside of the loop, and run the code again.
Now, this style of looping is actually much closer to languages like Java or C++. In Python it's quite rare to see code like this. You can just iterate over any sequence or collection, so there is no need to get the list of positions and retrieve elements out of a sequence at each iteration. It's expensive, needlessly expensive. Let's change the example into a more Pythonic form:
simple.for.3.py
surnames = ['Rivest', 'Shamir', 'Adleman'] for surname in surnames: print(surname)
Now that's something! It's practically English. The for
loop can iterate over the surnames
list, and it gives back each element in order at each interaction. Running this code will print the three surnames, one at a time. It's much easier to read, right?
What if you wanted to print the position as well though? Or what if you actually needed it for any reason? Should you go back to the range(len(...))
form? No. You can use the enumerate
built-in function, like this:
simple.for.4.py
surnames = ['Rivest', 'Shamir', 'Adleman'] for position, surname in enumerate(surnames): print(position, surname)
This code is very interesting as well. Notice that enumerate gives back a 2-tuple (position, surname)
at each iteration, but still, it's much more readable (and more efficient) than the range(len(...))
example. You can call enumerate
with a start
parameter, like enumerate(iterable, start)
, and it will start from start
, rather than 0
. Just another little thing that shows you how much thought has been given in designing Python so that it makes your life easy.
Using a for
loop it is possible to iterate over lists, tuples, and in general anything that in Python is called iterable. This is a very important concept, so let's talk about it a bit more.
According to the Python documentation, an iterable is:
"An object capable of returning its members one at a time. Examples of iterables include all sequence types (such as
list
,str
, and tuple) and some non-sequence types likedict
,file
objects, and objects of any classes you define with an__iter__()
or__getitem__()
method. Iterables can be used in afor
loop and in many other places where a sequence is needed (zip()
,map()
, ...). When an iterable object is passed as an argument to the built-in functioniter()
, it returns an iterator for the object. This iterator is good for one pass over the set of values. When using iterables, it is usually not necessary to calliter()
or deal with iterator objects yourself. Thefor
statement does that automatically for you, creating a temporary unnamed variable to hold the iterator for the duration of the loop."
Simply put, what happens when you write for k in sequence: ... body ...
, is that the for
loop asks sequence
for the next element, it gets something back, it calls that something k
, and then executes its body. Then, once again, the for
loop asks sequence
again for the next element, it calls it k
again, and executes the body again, and so on and so forth, until the sequence is exhausted. Empty sequences will result in zero executions of the body.
Some data structures, when iterated over, produce their elements in order, like lists, tuples, and strings, while some others don't, like sets and dictionaries.
Python gives us the ability to iterate over iterables, using a type of object called iterator. According to the official documentation, an iterator is:
"An object representing a stream of data. Repeated calls to the iterator's
__next__()
method (or passing it to the built-in functionnext()
) return successive items in the stream. When no more data are available aStopIteration
exception is raised instead. At this point, the iterator object is exhausted and any further calls to its__next__()
method just raiseStopIteration
again. Iterators are required to have an__iter__()
method that returns the iterator object itself so every iterator is also iterable and may be used in most places where other iterables are accepted. One notable exception is code which attempts multiple iteration passes. A container object (such as alist
) produces a fresh new iterator each time you pass it to theiter()
function or use it in afor
loop. Attempting this with an iterator will just return the same exhausted iterator object used in the previous iteration pass, making it appear like an empty container."
Don't worry if you don't fully understand all the preceding legalese, you will in due time. I put it here as a handy reference for the future.
In practice, the whole iterable/iterator mechanism is somewhat hidden behind the code. Unless you need to code your own iterable or iterator for some reason, you won't have to worry about this too much. But it's very important to understand how Python handles this key aspect of control flow because it will shape the way you will write your code.
Let's see another example of how to iterate over two sequences of the same length, in order to work on their respective elements in pairs. Say we have a list of people and a list of numbers representing the age of the people in the first list. We want to print a pair person/age on one line for all of them. Let's start with an example and let's refine it gradually.
multiple.sequences.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] for position in range(len(people)): person = people[position] age = ages[position] print(person, age)
By now, this code should be pretty straightforward for you to understand. We need to iterate over the list of positions (0, 1, 2, 3) because we want to retrieve elements from two different lists. Executing it we get the following:
$ python multiple.sequences.py Jonas 25 Julio 30 Mike 31 Mez 39
This code is both inefficient and not Pythonic. Inefficient because retrieving an element given the position can be an expensive operation, and we're doing it from scratch at each iteration. The mail man doesn't go back to the beginning of the road each time he delivers a letter, right? He moves from house to house. From one to the next one. Let's try to make it better using enumerate:
multiple.sequences.enumerate.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] for position, person in enumerate(people): age = ages[position] print(person, age)
Better, but still not perfect. And still a bit ugly. We're iterating properly on people
, but we're still fetching age
using positional indexing, which we want to lose as well. Well, no worries, Python gives you the zip
function, remember? Let's use it!
multiple.sequences.zip.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] for person, age in zip(people, ages): print(person, age)
Ah! So much better! Once again, compare the preceding code with the first example and admire Python's elegance. The reason I wanted to show this example is twofold. On the one hand, I wanted to give you an idea of how shorter the code in Python can be compared to other languages where the syntax doesn't allow you to iterate over sequences or collections as easily. And on the other hand, and much more importantly, notice that when the for
loop asks zip(sequenceA, sequenceB)
for the next element, it gets back a tuple
, not just a single object. It gets back a tuple
with as many elements as the number of sequences we feed to the zip
function. Let's expand a little on the previous example in two ways: using explicit and implicit assignment:
multiple.sequences.explicit.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] nationalities = ['Belgium', 'Spain', 'England', 'Bangladesh'] for person, age, nationality in zip(people, ages, nationalities): print(person, age, nationality)
In the preceding code, we added the nationalities list. Now that we feed three sequences to the zip
function, the for loop gets back a 3-tuple at each iteration. Notice that the position of the elements in the tuple respects the position of the sequences in the zip
call. Executing the code will yield the following result:
$ python multiple.sequences.explicit.py Jonas 25 Belgium Julio 30 Spain Mike 31 England Mez 39 Bangladesh
Sometimes, for reasons that may not be clear in a simple example like the preceding one, you may want to explode the tuple within the body of the for
loop. If that is your desire, it's perfectly possible to do so.
multiple.sequences.implicit.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] nationalities = ['Belgium', 'Spain', 'England', 'Bangladesh'] for data in zip(people, ages, nationalities): person, age, nationality = data print(person, age, nationality)
It's basically doing what the for
loop does automatically for you, but in some cases you may want to do it yourself. Here, the 3-tuple data
that comes from zip(...)
, is exploded within the body of the for
loop into three variables: person
, age
, and nationality
.
In the preceding pages, we saw the for
loop in action. It's incredibly useful when you need to loop over a sequence or a collection. The key point to keep in mind, when you need to be able to discriminate which looping construct to use, is that the for
loop rocks when you have to iterate over a finite amount of elements. It can be a huge amount, but still, something that at some point ends.
There are other cases though, when you just need to loop until some condition is satisfied, or even loop indefinitely until the application is stopped. Cases where we don't really have something to iterate on, and therefore the for
loop would be a poor choice. But fear not, for these cases Python provides us with the while
loop.
The while
loop is similar to the for
loop, in that they both loop and at each iteration they execute a body of instructions. What is different between them is that the while
loop doesn't loop over a sequence (it can, but you have to manually write the logic and it wouldn't make any sense, you would just want to use a for
loop), rather, it loops as long as a certain condition is satisfied. When the condition is no longer satisfied, the loop ends.
As usual, let's see an example which will clarify everything for us. We want to print the binary representation of a positive number. In order to do so, we repeatedly divide the number by two, collecting the remainder, and then produce the inverse of the list of remainders. Let me give you a small example using number 6, which is 110 in binary.
6 / 2 = 3 (remainder: 0) 3 / 2 = 1 (remainder: 1) 1 / 2 = 0 (remainder: 1) List of remainders: 0, 1, 1. Inverse is 1, 1, 0, which is also the binary representation of 6: 110
Let's write some code to calculate the binary representation for number 39: 1001112.
binary.py
n = 39 remainders = [] while n > 0: remainder = n % 2 # remainder of division by 2 remainders.append(remainder) # we keep track of remainders n //= 2 # we divide n by 2 # reassign the list to its reversed copy and print it remainders = remainders[::-1] print(remainders)
In the preceding code, I highlighted two things: n > 0
, which is the condition to keep looping, and remainders[::-1]
which is a nice and easy way to get the reversed version of a list (missing start
and end
parameters, step = -1
, produces the same list, from end
to start
, in reverse order). We can make the code a little shorter (and more Pythonic), by using the divmod
function, which is called with a number and a divisor, and returns a tuple with the result of the integer division and its remainder. For example, divmod(13, 5)
would return (2, 3)
, and indeed 5 * 2 + 3 = 13.
binary.2.py
n = 39
remainders = []
while n > 0:
n, remainder = divmod(n, 2)
remainders.append(remainder)
# reassign the list to its reversed copy and print it
remainders = remainders[::-1]
print(remainders)
In the preceding code, we have reassigned n to the result of the division by 2, and the remainder, in one single line.
Notice that the condition in a while
loop is a condition to continue looping. If it evaluates to True
, then the body is executed and then another evaluation follows, and so on, until the condition evaluates to False
. When that happens, the loop is exited immediately without executing its body.
Note
If the condition never evaluates to False
, the loop becomes a so called infinite loop. Infinite loops are used for example when polling from network devices: you ask the socket if there is any data, you do something with it if there is any, then you sleep for a small amount of time, and then you ask the socket again, over and over again, without ever stopping.
Having the ability to loop over a condition, or to loop indefinitely, is the reason why the for
loop alone is not enough, and therefore Python provides the while
loop.
Just for fun, let's adapt one of the examples (multiple.sequences.py
) using the while logic.
multiple.sequences.while.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] position = 0 while position < len(people): person = people[position] age = ages[position] print(person, age) position += 1
In the preceding code, I have highlighted the initialization, condition, and update of the variable position
, which makes it possible to simulate the equivalent for
loop code by handling the iteration variable manually. Everything that can be done with a for
loop can also be done with a while
loop, even though you can see there's a bit of boilerplate you have to go through in order to achieve the same result. The opposite is also true, but simulating a never ending while
loop using a for
loop requires some real trickery, so why would you do that? Use the right tool for the job, and 99.9% of the times you'll be fine.
So, to recap, use a for
loop when you need to iterate over one (or a combination of) iterable, and a while
loop when you need to loop according to a condition being satisfied or not. If you keep in mind the difference between the two purposes, you will never choose the wrong looping construct.
Let's now see how to alter the normal flow of a loop.
According to the task at hand, sometimes you will need to alter the regular flow of a loop. You can either skip a single iteration (as many times you want), or you can break out of the loop entirely. A common use case for skipping iterations is for example when you're iterating over a list of items and you need to work on each of them only if some condition is verified. On the other hand, if you're iterating over a collection of items, and you have found one of them that satisfies some need you have, you may decide not to continue the loop entirely and therefore break out of it. There are countless possible scenarios, so it's better to see a couple of examples.
Let's say you want to apply a 20% discount to all products in a basket list for those which have an expiration date of today. The way you achieve this is to use the continue statement, which tells the looping construct (for
or while
) to immediately stop execution of the body and go to the next iteration, if any. This example will take us a little deeper down the rabbit whole, so be ready to jump.
discount.py
from datetime import date, timedelta today = date.today() tomorrow = today + timedelta(days=1) # today + 1 day is tomorrow products = [ {'sku': '1', 'expiration_date': today, 'price': 100.0}, {'sku': '2', 'expiration_date': tomorrow, 'price': 50}, {'sku': '3', 'expiration_date': today, 'price': 20}, ] for product in products: if product['expiration_date'] != today: continue product['price'] *= 0.8 # equivalent to applying 20% discount print( 'Price for sku', product['sku'], 'is now', product['price'])
You see we start by importing the date
and timedelta
objects, then we set up our products. Those with sku 1
and 3
have an expiration date of today
, which means we want to apply 20% discount on them. We loop over each product
and we inspect the expiration date. If it is not (inequality operator, !=
) today
, we don't want to execute the rest of the body suite, so we continue
.
Notice that is not important where in the body suite you place the continue
statement (you can even use it more than once). When you reach it, execution stops and goes back to the next iteration. If we run the discount.py
module, this is the output:
$ python discount.py Price for sku 1 is now 80.0 Price for sku 3 is now 16.0
Which shows you that the last two lines of the body haven't been executed for sku number 2.
Let's now see an example of breaking out of a loop. Say we want to tell if at least any of the elements in a list evaluates to True
when fed to the bool
function. Given that we need to know if there is at least one, when we find it we don't need to keep scanning the list any further. In Python code, this translates to using the break statement. Let's write this down into code:
any.py
items = [0, None, 0.0, True, 0, 7] # True and 7 evaluate to True found = False # this is called "flag" for item in items: print('scanning item', item) if item: found = True # we update the flag break if found: # we inspect the flag print('At least one item evaluates to True') else: print('All items evaluate to False')
The preceding code is such a common pattern in programming, you will see it a lot. When you inspect items this way, basically what you do is to set up a flag
variable, then start the inspection. If you find one element that matches your criteria (in this example, that evaluates to True
), then you update the flag and stop iterating. After iteration, you inspect the flag and take action accordingly. Execution yields:
$ python any.py scanning item 0 scanning item None scanning item 0.0 scanning item True At least one item evaluates to True
See how execution stopped after True
was found?
The break
statement acts exactly like the continue
one, in that it stops executing the body of the loop immediately, but also, prevents any other iteration to run, effectively breaking out of the loop.
The continue
and break
statements can be used together with no limitation in their number, both in the for
and while
looping constructs.
One of the features I've seen only in the Python language is the ability to have else
clauses after while
and for
loops. It's very rarely used, but it's definitely nice to have. In short, you can have an else
suite after a for
or while
loop. If the loop ends normally, because of exhaustion of the iterator (for
loop) or because the condition is finally not met (while
loop), then the else
suite (if present) is executed. In case execution is interrupted by a break
statement, the else
clause is not executed. Let's take an example of a for
loop that iterates over a group of items, looking for one that would match some condition. In case we don't find at least one that satisfies the condition, we want to raise an exception. This means we want to arrest the regular execution of the program and signal that there was an error, or exception, that we cannot deal with. Exceptions will be the subject of Chapter 7, Testing, Profiling, and Dealing with Exceptions, so don't worry if you don't fully understand them now. Just bear in mind that they will alter the regular flow of the code. Let me now show you two examples that do exactly the same thing, but one of them is using the special for
... else
syntax. Say that we want to find among a collection of people one that could drive a car.
for.no.else.py
class DriverException(Exception): pass people = [('James', 17), ('Kirk', 9), ('Lars', 13), ('Robert', 8)] driver = None for person, age in people: if age >= 18: driver = (person, age) break if driver is None: raise DriverException('Driver not found.')
Notice the flag
pattern again. We set driver to be None
, then if we find one we update the driver
flag, and then, at the end of the loop, we inspect it to see if one was found. I kind of have the feeling that those kids would drive a very metallic car, but anyway, notice that if a driver is not found, a DriverException
is raised, signaling the program that execution cannot continue (we're lacking the driver).
The same functionality can be rewritten a bit more elegantly using the following code:
for.else.py
class DriverException(Exception): pass people = [('James', 17), ('Kirk', 9), ('Lars', 13), ('Robert', 8)] for person, age in people: if age >= 18: driver = (person, age) break else: raise DriverException('Driver not found.')
Notice that we aren't forced to use the flag
pattern any more. The exception is raised as part of the for
loop logic, which makes good sense because the for
loop is checking on some condition. All we need is to set up a driver
object in case we find one, because the rest of the code is going to use that information somewhere. Notice the code is shorter and more elegant, because the logic is now correctly grouped together where it belongs.
The for loop
The for
loop is used when looping over a sequence, like a list, tuple, or a collection of objects. Let's start with a simple example that is more like C++ style, and then let's gradually see how to achieve the same results in Python (you'll love Python's syntax).
simple.for.py
for number in [0, 1, 2, 3, 4]: print(number)
This simple snippet of code, when executed, prints all numbers from 0 to 4. The for
loop is fed the list [0, 1, 2, 3, 4]
and at each iteration, number
is given a value from the sequence (which is iterated sequentially, in order), then the body of the loop is executed (the print line). number
changes at every iteration, according to which value is coming next from the sequence. When the sequence is exhausted, the for
loop terminates, and the execution of the code resumes normally with the code after the loop.
Sometimes we need to iterate over a range of numbers, and it would be quite unpleasant to have to do so by hardcoding the list somewhere. In such cases, the range
function comes to the rescue. Let's see the equivalent of the previous snippet of code:
simple.for.py
for number in range(5): print(number)
The range function is used extensively in Python programs when it comes to creating sequences: you can call it by passing one value, which acts as stop
(counting from 0), or you can pass two values (start
and stop
), or even three (start
, stop
, and step
). Check out the following example:
>>> list(range(10)) # one value: from 0 to value (excluded) [0, 1, 2, 3, 4, 5, 6, 7, 8, 9] >>> list(range(3, 8)) # two values: from start to stop (excluded) [3, 4, 5, 6, 7] >>> list(range(-10, 10, 4)) # three values: step is added [-10, -6, -2, 2, 6]
For the moment, ignore that we need to wrap range(...)
within a list
. The range
object is a little bit special, but in this case we're just interested in understanding what are the values it will return to us. You see that the deal is the same with slicing: start
is included, stop
excluded, and optionally you can add a step
parameter, which by default is 1.
Try modifying the parameters of the range()
call in our simple.for.py
code and see what it prints, get comfortable with it.
Now we have all the tools to iterate over a sequence, so let's build on that example:
simple.for.2.py
surnames = ['Rivest', 'Shamir', 'Adleman'] for position in range(len(surnames)): print(position, surnames[position])
The preceding code adds a little bit of complexity to the game. Execution will show this result:
$ python simple.for.2.py 0 Rivest 1 Shamir 2 Adleman
Let's use the inside-out technique to break it down, ok? We start from the innermost part of what we're trying to understand, and we expand outwards. So, len(surnames)
is the length of the surnames
list: 3
. Therefore, range(len(surnames))
is actually transformed into range(3)
. This gives us the range [0, 3), which is basically a sequence (0, 1, 2)
. This means that the for
loop will run three iterations. In the first one, position
will take value 0
, while in the second one, it will take value 1
, and finally value 2
in the third and last iteration. What is (0, 1, 2)
, if not the possible indexing positions for the surnames
list? At position 0
we find 'Rivest'
, at position 1
, 'Shamir'
, and at position 2
, 'Adleman'
. If you are curious about what these three men created together, change print(position, surnames[position])
to print(surnames[position][0], end='')
add a final print()
outside of the loop, and run the code again.
Now, this style of looping is actually much closer to languages like Java or C++. In Python it's quite rare to see code like this. You can just iterate over any sequence or collection, so there is no need to get the list of positions and retrieve elements out of a sequence at each iteration. It's expensive, needlessly expensive. Let's change the example into a more Pythonic form:
simple.for.3.py
surnames = ['Rivest', 'Shamir', 'Adleman'] for surname in surnames: print(surname)
Now that's something! It's practically English. The for
loop can iterate over the surnames
list, and it gives back each element in order at each interaction. Running this code will print the three surnames, one at a time. It's much easier to read, right?
What if you wanted to print the position as well though? Or what if you actually needed it for any reason? Should you go back to the range(len(...))
form? No. You can use the enumerate
built-in function, like this:
simple.for.4.py
surnames = ['Rivest', 'Shamir', 'Adleman'] for position, surname in enumerate(surnames): print(position, surname)
This code is very interesting as well. Notice that enumerate gives back a 2-tuple (position, surname)
at each iteration, but still, it's much more readable (and more efficient) than the range(len(...))
example. You can call enumerate
with a start
parameter, like enumerate(iterable, start)
, and it will start from start
, rather than 0
. Just another little thing that shows you how much thought has been given in designing Python so that it makes your life easy.
Using a for
loop it is possible to iterate over lists, tuples, and in general anything that in Python is called iterable. This is a very important concept, so let's talk about it a bit more.
According to the Python documentation, an iterable is:
"An object capable of returning its members one at a time. Examples of iterables include all sequence types (such as
list
,str
, and tuple) and some non-sequence types likedict
,file
objects, and objects of any classes you define with an__iter__()
or__getitem__()
method. Iterables can be used in afor
loop and in many other places where a sequence is needed (zip()
,map()
, ...). When an iterable object is passed as an argument to the built-in functioniter()
, it returns an iterator for the object. This iterator is good for one pass over the set of values. When using iterables, it is usually not necessary to calliter()
or deal with iterator objects yourself. Thefor
statement does that automatically for you, creating a temporary unnamed variable to hold the iterator for the duration of the loop."
Simply put, what happens when you write for k in sequence: ... body ...
, is that the for
loop asks sequence
for the next element, it gets something back, it calls that something k
, and then executes its body. Then, once again, the for
loop asks sequence
again for the next element, it calls it k
again, and executes the body again, and so on and so forth, until the sequence is exhausted. Empty sequences will result in zero executions of the body.
Some data structures, when iterated over, produce their elements in order, like lists, tuples, and strings, while some others don't, like sets and dictionaries.
Python gives us the ability to iterate over iterables, using a type of object called iterator. According to the official documentation, an iterator is:
"An object representing a stream of data. Repeated calls to the iterator's
__next__()
method (or passing it to the built-in functionnext()
) return successive items in the stream. When no more data are available aStopIteration
exception is raised instead. At this point, the iterator object is exhausted and any further calls to its__next__()
method just raiseStopIteration
again. Iterators are required to have an__iter__()
method that returns the iterator object itself so every iterator is also iterable and may be used in most places where other iterables are accepted. One notable exception is code which attempts multiple iteration passes. A container object (such as alist
) produces a fresh new iterator each time you pass it to theiter()
function or use it in afor
loop. Attempting this with an iterator will just return the same exhausted iterator object used in the previous iteration pass, making it appear like an empty container."
Don't worry if you don't fully understand all the preceding legalese, you will in due time. I put it here as a handy reference for the future.
In practice, the whole iterable/iterator mechanism is somewhat hidden behind the code. Unless you need to code your own iterable or iterator for some reason, you won't have to worry about this too much. But it's very important to understand how Python handles this key aspect of control flow because it will shape the way you will write your code.
Let's see another example of how to iterate over two sequences of the same length, in order to work on their respective elements in pairs. Say we have a list of people and a list of numbers representing the age of the people in the first list. We want to print a pair person/age on one line for all of them. Let's start with an example and let's refine it gradually.
multiple.sequences.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] for position in range(len(people)): person = people[position] age = ages[position] print(person, age)
By now, this code should be pretty straightforward for you to understand. We need to iterate over the list of positions (0, 1, 2, 3) because we want to retrieve elements from two different lists. Executing it we get the following:
$ python multiple.sequences.py Jonas 25 Julio 30 Mike 31 Mez 39
This code is both inefficient and not Pythonic. Inefficient because retrieving an element given the position can be an expensive operation, and we're doing it from scratch at each iteration. The mail man doesn't go back to the beginning of the road each time he delivers a letter, right? He moves from house to house. From one to the next one. Let's try to make it better using enumerate:
multiple.sequences.enumerate.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] for position, person in enumerate(people): age = ages[position] print(person, age)
Better, but still not perfect. And still a bit ugly. We're iterating properly on people
, but we're still fetching age
using positional indexing, which we want to lose as well. Well, no worries, Python gives you the zip
function, remember? Let's use it!
multiple.sequences.zip.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] for person, age in zip(people, ages): print(person, age)
Ah! So much better! Once again, compare the preceding code with the first example and admire Python's elegance. The reason I wanted to show this example is twofold. On the one hand, I wanted to give you an idea of how shorter the code in Python can be compared to other languages where the syntax doesn't allow you to iterate over sequences or collections as easily. And on the other hand, and much more importantly, notice that when the for
loop asks zip(sequenceA, sequenceB)
for the next element, it gets back a tuple
, not just a single object. It gets back a tuple
with as many elements as the number of sequences we feed to the zip
function. Let's expand a little on the previous example in two ways: using explicit and implicit assignment:
multiple.sequences.explicit.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] nationalities = ['Belgium', 'Spain', 'England', 'Bangladesh'] for person, age, nationality in zip(people, ages, nationalities): print(person, age, nationality)
In the preceding code, we added the nationalities list. Now that we feed three sequences to the zip
function, the for loop gets back a 3-tuple at each iteration. Notice that the position of the elements in the tuple respects the position of the sequences in the zip
call. Executing the code will yield the following result:
$ python multiple.sequences.explicit.py Jonas 25 Belgium Julio 30 Spain Mike 31 England Mez 39 Bangladesh
Sometimes, for reasons that may not be clear in a simple example like the preceding one, you may want to explode the tuple within the body of the for
loop. If that is your desire, it's perfectly possible to do so.
multiple.sequences.implicit.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] nationalities = ['Belgium', 'Spain', 'England', 'Bangladesh'] for data in zip(people, ages, nationalities): person, age, nationality = data print(person, age, nationality)
It's basically doing what the for
loop does automatically for you, but in some cases you may want to do it yourself. Here, the 3-tuple data
that comes from zip(...)
, is exploded within the body of the for
loop into three variables: person
, age
, and nationality
.
In the preceding pages, we saw the for
loop in action. It's incredibly useful when you need to loop over a sequence or a collection. The key point to keep in mind, when you need to be able to discriminate which looping construct to use, is that the for
loop rocks when you have to iterate over a finite amount of elements. It can be a huge amount, but still, something that at some point ends.
There are other cases though, when you just need to loop until some condition is satisfied, or even loop indefinitely until the application is stopped. Cases where we don't really have something to iterate on, and therefore the for
loop would be a poor choice. But fear not, for these cases Python provides us with the while
loop.
The while
loop is similar to the for
loop, in that they both loop and at each iteration they execute a body of instructions. What is different between them is that the while
loop doesn't loop over a sequence (it can, but you have to manually write the logic and it wouldn't make any sense, you would just want to use a for
loop), rather, it loops as long as a certain condition is satisfied. When the condition is no longer satisfied, the loop ends.
As usual, let's see an example which will clarify everything for us. We want to print the binary representation of a positive number. In order to do so, we repeatedly divide the number by two, collecting the remainder, and then produce the inverse of the list of remainders. Let me give you a small example using number 6, which is 110 in binary.
6 / 2 = 3 (remainder: 0) 3 / 2 = 1 (remainder: 1) 1 / 2 = 0 (remainder: 1) List of remainders: 0, 1, 1. Inverse is 1, 1, 0, which is also the binary representation of 6: 110
Let's write some code to calculate the binary representation for number 39: 1001112.
binary.py
n = 39 remainders = [] while n > 0: remainder = n % 2 # remainder of division by 2 remainders.append(remainder) # we keep track of remainders n //= 2 # we divide n by 2 # reassign the list to its reversed copy and print it remainders = remainders[::-1] print(remainders)
In the preceding code, I highlighted two things: n > 0
, which is the condition to keep looping, and remainders[::-1]
which is a nice and easy way to get the reversed version of a list (missing start
and end
parameters, step = -1
, produces the same list, from end
to start
, in reverse order). We can make the code a little shorter (and more Pythonic), by using the divmod
function, which is called with a number and a divisor, and returns a tuple with the result of the integer division and its remainder. For example, divmod(13, 5)
would return (2, 3)
, and indeed 5 * 2 + 3 = 13.
binary.2.py
n = 39
remainders = []
while n > 0:
n, remainder = divmod(n, 2)
remainders.append(remainder)
# reassign the list to its reversed copy and print it
remainders = remainders[::-1]
print(remainders)
In the preceding code, we have reassigned n to the result of the division by 2, and the remainder, in one single line.
Notice that the condition in a while
loop is a condition to continue looping. If it evaluates to True
, then the body is executed and then another evaluation follows, and so on, until the condition evaluates to False
. When that happens, the loop is exited immediately without executing its body.
Note
If the condition never evaluates to False
, the loop becomes a so called infinite loop. Infinite loops are used for example when polling from network devices: you ask the socket if there is any data, you do something with it if there is any, then you sleep for a small amount of time, and then you ask the socket again, over and over again, without ever stopping.
Having the ability to loop over a condition, or to loop indefinitely, is the reason why the for
loop alone is not enough, and therefore Python provides the while
loop.
Just for fun, let's adapt one of the examples (multiple.sequences.py
) using the while logic.
multiple.sequences.while.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] position = 0 while position < len(people): person = people[position] age = ages[position] print(person, age) position += 1
In the preceding code, I have highlighted the initialization, condition, and update of the variable position
, which makes it possible to simulate the equivalent for
loop code by handling the iteration variable manually. Everything that can be done with a for
loop can also be done with a while
loop, even though you can see there's a bit of boilerplate you have to go through in order to achieve the same result. The opposite is also true, but simulating a never ending while
loop using a for
loop requires some real trickery, so why would you do that? Use the right tool for the job, and 99.9% of the times you'll be fine.
So, to recap, use a for
loop when you need to iterate over one (or a combination of) iterable, and a while
loop when you need to loop according to a condition being satisfied or not. If you keep in mind the difference between the two purposes, you will never choose the wrong looping construct.
Let's now see how to alter the normal flow of a loop.
According to the task at hand, sometimes you will need to alter the regular flow of a loop. You can either skip a single iteration (as many times you want), or you can break out of the loop entirely. A common use case for skipping iterations is for example when you're iterating over a list of items and you need to work on each of them only if some condition is verified. On the other hand, if you're iterating over a collection of items, and you have found one of them that satisfies some need you have, you may decide not to continue the loop entirely and therefore break out of it. There are countless possible scenarios, so it's better to see a couple of examples.
Let's say you want to apply a 20% discount to all products in a basket list for those which have an expiration date of today. The way you achieve this is to use the continue statement, which tells the looping construct (for
or while
) to immediately stop execution of the body and go to the next iteration, if any. This example will take us a little deeper down the rabbit whole, so be ready to jump.
discount.py
from datetime import date, timedelta today = date.today() tomorrow = today + timedelta(days=1) # today + 1 day is tomorrow products = [ {'sku': '1', 'expiration_date': today, 'price': 100.0}, {'sku': '2', 'expiration_date': tomorrow, 'price': 50}, {'sku': '3', 'expiration_date': today, 'price': 20}, ] for product in products: if product['expiration_date'] != today: continue product['price'] *= 0.8 # equivalent to applying 20% discount print( 'Price for sku', product['sku'], 'is now', product['price'])
You see we start by importing the date
and timedelta
objects, then we set up our products. Those with sku 1
and 3
have an expiration date of today
, which means we want to apply 20% discount on them. We loop over each product
and we inspect the expiration date. If it is not (inequality operator, !=
) today
, we don't want to execute the rest of the body suite, so we continue
.
Notice that is not important where in the body suite you place the continue
statement (you can even use it more than once). When you reach it, execution stops and goes back to the next iteration. If we run the discount.py
module, this is the output:
$ python discount.py Price for sku 1 is now 80.0 Price for sku 3 is now 16.0
Which shows you that the last two lines of the body haven't been executed for sku number 2.
Let's now see an example of breaking out of a loop. Say we want to tell if at least any of the elements in a list evaluates to True
when fed to the bool
function. Given that we need to know if there is at least one, when we find it we don't need to keep scanning the list any further. In Python code, this translates to using the break statement. Let's write this down into code:
any.py
items = [0, None, 0.0, True, 0, 7] # True and 7 evaluate to True found = False # this is called "flag" for item in items: print('scanning item', item) if item: found = True # we update the flag break if found: # we inspect the flag print('At least one item evaluates to True') else: print('All items evaluate to False')
The preceding code is such a common pattern in programming, you will see it a lot. When you inspect items this way, basically what you do is to set up a flag
variable, then start the inspection. If you find one element that matches your criteria (in this example, that evaluates to True
), then you update the flag and stop iterating. After iteration, you inspect the flag and take action accordingly. Execution yields:
$ python any.py scanning item 0 scanning item None scanning item 0.0 scanning item True At least one item evaluates to True
See how execution stopped after True
was found?
The break
statement acts exactly like the continue
one, in that it stops executing the body of the loop immediately, but also, prevents any other iteration to run, effectively breaking out of the loop.
The continue
and break
statements can be used together with no limitation in their number, both in the for
and while
looping constructs.
One of the features I've seen only in the Python language is the ability to have else
clauses after while
and for
loops. It's very rarely used, but it's definitely nice to have. In short, you can have an else
suite after a for
or while
loop. If the loop ends normally, because of exhaustion of the iterator (for
loop) or because the condition is finally not met (while
loop), then the else
suite (if present) is executed. In case execution is interrupted by a break
statement, the else
clause is not executed. Let's take an example of a for
loop that iterates over a group of items, looking for one that would match some condition. In case we don't find at least one that satisfies the condition, we want to raise an exception. This means we want to arrest the regular execution of the program and signal that there was an error, or exception, that we cannot deal with. Exceptions will be the subject of Chapter 7, Testing, Profiling, and Dealing with Exceptions, so don't worry if you don't fully understand them now. Just bear in mind that they will alter the regular flow of the code. Let me now show you two examples that do exactly the same thing, but one of them is using the special for
... else
syntax. Say that we want to find among a collection of people one that could drive a car.
for.no.else.py
class DriverException(Exception): pass people = [('James', 17), ('Kirk', 9), ('Lars', 13), ('Robert', 8)] driver = None for person, age in people: if age >= 18: driver = (person, age) break if driver is None: raise DriverException('Driver not found.')
Notice the flag
pattern again. We set driver to be None
, then if we find one we update the driver
flag, and then, at the end of the loop, we inspect it to see if one was found. I kind of have the feeling that those kids would drive a very metallic car, but anyway, notice that if a driver is not found, a DriverException
is raised, signaling the program that execution cannot continue (we're lacking the driver).
The same functionality can be rewritten a bit more elegantly using the following code:
for.else.py
class DriverException(Exception): pass people = [('James', 17), ('Kirk', 9), ('Lars', 13), ('Robert', 8)] for person, age in people: if age >= 18: driver = (person, age) break else: raise DriverException('Driver not found.')
Notice that we aren't forced to use the flag
pattern any more. The exception is raised as part of the for
loop logic, which makes good sense because the for
loop is checking on some condition. All we need is to set up a driver
object in case we find one, because the rest of the code is going to use that information somewhere. Notice the code is shorter and more elegant, because the logic is now correctly grouped together where it belongs.
Iterating over a range
Sometimes we need to iterate over a range of numbers, and it would be quite unpleasant to have to do so by hardcoding the list somewhere. In such cases, the range
function comes to the rescue. Let's see the equivalent of the previous snippet of code:
simple.for.py
for number in range(5): print(number)
The range function is used extensively in Python programs when it comes to creating sequences: you can call it by passing one value, which acts as stop
(counting from 0), or you can pass two values (start
and stop
), or even three (start
, stop
, and step
). Check out the following example:
>>> list(range(10)) # one value: from 0 to value (excluded) [0, 1, 2, 3, 4, 5, 6, 7, 8, 9] >>> list(range(3, 8)) # two values: from start to stop (excluded) [3, 4, 5, 6, 7] >>> list(range(-10, 10, 4)) # three values: step is added [-10, -6, -2, 2, 6]
For the moment, ignore that we need to wrap range(...)
within a list
. The range
object is a little bit special, but in this case we're just interested in understanding what are the values it will return to us. You see that the deal is the same with slicing: start
is included, stop
excluded, and optionally you can add a step
parameter, which by default is 1.
Try modifying the parameters of the range()
call in our simple.for.py
code and see what it prints, get comfortable with it.
Now we have all the tools to iterate over a sequence, so let's build on that example:
simple.for.2.py
surnames = ['Rivest', 'Shamir', 'Adleman'] for position in range(len(surnames)): print(position, surnames[position])
The preceding code adds a little bit of complexity to the game. Execution will show this result:
$ python simple.for.2.py 0 Rivest 1 Shamir 2 Adleman
Let's use the inside-out technique to break it down, ok? We start from the innermost part of what we're trying to understand, and we expand outwards. So, len(surnames)
is the length of the surnames
list: 3
. Therefore, range(len(surnames))
is actually transformed into range(3)
. This gives us the range [0, 3), which is basically a sequence (0, 1, 2)
. This means that the for
loop will run three iterations. In the first one, position
will take value 0
, while in the second one, it will take value 1
, and finally value 2
in the third and last iteration. What is (0, 1, 2)
, if not the possible indexing positions for the surnames
list? At position 0
we find 'Rivest'
, at position 1
, 'Shamir'
, and at position 2
, 'Adleman'
. If you are curious about what these three men created together, change print(position, surnames[position])
to print(surnames[position][0], end='')
add a final print()
outside of the loop, and run the code again.
Now, this style of looping is actually much closer to languages like Java or C++. In Python it's quite rare to see code like this. You can just iterate over any sequence or collection, so there is no need to get the list of positions and retrieve elements out of a sequence at each iteration. It's expensive, needlessly expensive. Let's change the example into a more Pythonic form:
simple.for.3.py
surnames = ['Rivest', 'Shamir', 'Adleman'] for surname in surnames: print(surname)
Now that's something! It's practically English. The for
loop can iterate over the surnames
list, and it gives back each element in order at each interaction. Running this code will print the three surnames, one at a time. It's much easier to read, right?
What if you wanted to print the position as well though? Or what if you actually needed it for any reason? Should you go back to the range(len(...))
form? No. You can use the enumerate
built-in function, like this:
simple.for.4.py
surnames = ['Rivest', 'Shamir', 'Adleman'] for position, surname in enumerate(surnames): print(position, surname)
This code is very interesting as well. Notice that enumerate gives back a 2-tuple (position, surname)
at each iteration, but still, it's much more readable (and more efficient) than the range(len(...))
example. You can call enumerate
with a start
parameter, like enumerate(iterable, start)
, and it will start from start
, rather than 0
. Just another little thing that shows you how much thought has been given in designing Python so that it makes your life easy.
Using a for
loop it is possible to iterate over lists, tuples, and in general anything that in Python is called iterable. This is a very important concept, so let's talk about it a bit more.
According to the Python documentation, an iterable is:
"An object capable of returning its members one at a time. Examples of iterables include all sequence types (such as
list
,str
, and tuple) and some non-sequence types likedict
,file
objects, and objects of any classes you define with an__iter__()
or__getitem__()
method. Iterables can be used in afor
loop and in many other places where a sequence is needed (zip()
,map()
, ...). When an iterable object is passed as an argument to the built-in functioniter()
, it returns an iterator for the object. This iterator is good for one pass over the set of values. When using iterables, it is usually not necessary to calliter()
or deal with iterator objects yourself. Thefor
statement does that automatically for you, creating a temporary unnamed variable to hold the iterator for the duration of the loop."
Simply put, what happens when you write for k in sequence: ... body ...
, is that the for
loop asks sequence
for the next element, it gets something back, it calls that something k
, and then executes its body. Then, once again, the for
loop asks sequence
again for the next element, it calls it k
again, and executes the body again, and so on and so forth, until the sequence is exhausted. Empty sequences will result in zero executions of the body.
Some data structures, when iterated over, produce their elements in order, like lists, tuples, and strings, while some others don't, like sets and dictionaries.
Python gives us the ability to iterate over iterables, using a type of object called iterator. According to the official documentation, an iterator is:
"An object representing a stream of data. Repeated calls to the iterator's
__next__()
method (or passing it to the built-in functionnext()
) return successive items in the stream. When no more data are available aStopIteration
exception is raised instead. At this point, the iterator object is exhausted and any further calls to its__next__()
method just raiseStopIteration
again. Iterators are required to have an__iter__()
method that returns the iterator object itself so every iterator is also iterable and may be used in most places where other iterables are accepted. One notable exception is code which attempts multiple iteration passes. A container object (such as alist
) produces a fresh new iterator each time you pass it to theiter()
function or use it in afor
loop. Attempting this with an iterator will just return the same exhausted iterator object used in the previous iteration pass, making it appear like an empty container."
Don't worry if you don't fully understand all the preceding legalese, you will in due time. I put it here as a handy reference for the future.
In practice, the whole iterable/iterator mechanism is somewhat hidden behind the code. Unless you need to code your own iterable or iterator for some reason, you won't have to worry about this too much. But it's very important to understand how Python handles this key aspect of control flow because it will shape the way you will write your code.
Let's see another example of how to iterate over two sequences of the same length, in order to work on their respective elements in pairs. Say we have a list of people and a list of numbers representing the age of the people in the first list. We want to print a pair person/age on one line for all of them. Let's start with an example and let's refine it gradually.
multiple.sequences.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] for position in range(len(people)): person = people[position] age = ages[position] print(person, age)
By now, this code should be pretty straightforward for you to understand. We need to iterate over the list of positions (0, 1, 2, 3) because we want to retrieve elements from two different lists. Executing it we get the following:
$ python multiple.sequences.py Jonas 25 Julio 30 Mike 31 Mez 39
This code is both inefficient and not Pythonic. Inefficient because retrieving an element given the position can be an expensive operation, and we're doing it from scratch at each iteration. The mail man doesn't go back to the beginning of the road each time he delivers a letter, right? He moves from house to house. From one to the next one. Let's try to make it better using enumerate:
multiple.sequences.enumerate.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] for position, person in enumerate(people): age = ages[position] print(person, age)
Better, but still not perfect. And still a bit ugly. We're iterating properly on people
, but we're still fetching age
using positional indexing, which we want to lose as well. Well, no worries, Python gives you the zip
function, remember? Let's use it!
multiple.sequences.zip.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] for person, age in zip(people, ages): print(person, age)
Ah! So much better! Once again, compare the preceding code with the first example and admire Python's elegance. The reason I wanted to show this example is twofold. On the one hand, I wanted to give you an idea of how shorter the code in Python can be compared to other languages where the syntax doesn't allow you to iterate over sequences or collections as easily. And on the other hand, and much more importantly, notice that when the for
loop asks zip(sequenceA, sequenceB)
for the next element, it gets back a tuple
, not just a single object. It gets back a tuple
with as many elements as the number of sequences we feed to the zip
function. Let's expand a little on the previous example in two ways: using explicit and implicit assignment:
multiple.sequences.explicit.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] nationalities = ['Belgium', 'Spain', 'England', 'Bangladesh'] for person, age, nationality in zip(people, ages, nationalities): print(person, age, nationality)
In the preceding code, we added the nationalities list. Now that we feed three sequences to the zip
function, the for loop gets back a 3-tuple at each iteration. Notice that the position of the elements in the tuple respects the position of the sequences in the zip
call. Executing the code will yield the following result:
$ python multiple.sequences.explicit.py Jonas 25 Belgium Julio 30 Spain Mike 31 England Mez 39 Bangladesh
Sometimes, for reasons that may not be clear in a simple example like the preceding one, you may want to explode the tuple within the body of the for
loop. If that is your desire, it's perfectly possible to do so.
multiple.sequences.implicit.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] nationalities = ['Belgium', 'Spain', 'England', 'Bangladesh'] for data in zip(people, ages, nationalities): person, age, nationality = data print(person, age, nationality)
It's basically doing what the for
loop does automatically for you, but in some cases you may want to do it yourself. Here, the 3-tuple data
that comes from zip(...)
, is exploded within the body of the for
loop into three variables: person
, age
, and nationality
.
In the preceding pages, we saw the for
loop in action. It's incredibly useful when you need to loop over a sequence or a collection. The key point to keep in mind, when you need to be able to discriminate which looping construct to use, is that the for
loop rocks when you have to iterate over a finite amount of elements. It can be a huge amount, but still, something that at some point ends.
There are other cases though, when you just need to loop until some condition is satisfied, or even loop indefinitely until the application is stopped. Cases where we don't really have something to iterate on, and therefore the for
loop would be a poor choice. But fear not, for these cases Python provides us with the while
loop.
The while
loop is similar to the for
loop, in that they both loop and at each iteration they execute a body of instructions. What is different between them is that the while
loop doesn't loop over a sequence (it can, but you have to manually write the logic and it wouldn't make any sense, you would just want to use a for
loop), rather, it loops as long as a certain condition is satisfied. When the condition is no longer satisfied, the loop ends.
As usual, let's see an example which will clarify everything for us. We want to print the binary representation of a positive number. In order to do so, we repeatedly divide the number by two, collecting the remainder, and then produce the inverse of the list of remainders. Let me give you a small example using number 6, which is 110 in binary.
6 / 2 = 3 (remainder: 0) 3 / 2 = 1 (remainder: 1) 1 / 2 = 0 (remainder: 1) List of remainders: 0, 1, 1. Inverse is 1, 1, 0, which is also the binary representation of 6: 110
Let's write some code to calculate the binary representation for number 39: 1001112.
binary.py
n = 39 remainders = [] while n > 0: remainder = n % 2 # remainder of division by 2 remainders.append(remainder) # we keep track of remainders n //= 2 # we divide n by 2 # reassign the list to its reversed copy and print it remainders = remainders[::-1] print(remainders)
In the preceding code, I highlighted two things: n > 0
, which is the condition to keep looping, and remainders[::-1]
which is a nice and easy way to get the reversed version of a list (missing start
and end
parameters, step = -1
, produces the same list, from end
to start
, in reverse order). We can make the code a little shorter (and more Pythonic), by using the divmod
function, which is called with a number and a divisor, and returns a tuple with the result of the integer division and its remainder. For example, divmod(13, 5)
would return (2, 3)
, and indeed 5 * 2 + 3 = 13.
binary.2.py
n = 39
remainders = []
while n > 0:
n, remainder = divmod(n, 2)
remainders.append(remainder)
# reassign the list to its reversed copy and print it
remainders = remainders[::-1]
print(remainders)
In the preceding code, we have reassigned n to the result of the division by 2, and the remainder, in one single line.
Notice that the condition in a while
loop is a condition to continue looping. If it evaluates to True
, then the body is executed and then another evaluation follows, and so on, until the condition evaluates to False
. When that happens, the loop is exited immediately without executing its body.
Note
If the condition never evaluates to False
, the loop becomes a so called infinite loop. Infinite loops are used for example when polling from network devices: you ask the socket if there is any data, you do something with it if there is any, then you sleep for a small amount of time, and then you ask the socket again, over and over again, without ever stopping.
Having the ability to loop over a condition, or to loop indefinitely, is the reason why the for
loop alone is not enough, and therefore Python provides the while
loop.
Just for fun, let's adapt one of the examples (multiple.sequences.py
) using the while logic.
multiple.sequences.while.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] position = 0 while position < len(people): person = people[position] age = ages[position] print(person, age) position += 1
In the preceding code, I have highlighted the initialization, condition, and update of the variable position
, which makes it possible to simulate the equivalent for
loop code by handling the iteration variable manually. Everything that can be done with a for
loop can also be done with a while
loop, even though you can see there's a bit of boilerplate you have to go through in order to achieve the same result. The opposite is also true, but simulating a never ending while
loop using a for
loop requires some real trickery, so why would you do that? Use the right tool for the job, and 99.9% of the times you'll be fine.
So, to recap, use a for
loop when you need to iterate over one (or a combination of) iterable, and a while
loop when you need to loop according to a condition being satisfied or not. If you keep in mind the difference between the two purposes, you will never choose the wrong looping construct.
Let's now see how to alter the normal flow of a loop.
According to the task at hand, sometimes you will need to alter the regular flow of a loop. You can either skip a single iteration (as many times you want), or you can break out of the loop entirely. A common use case for skipping iterations is for example when you're iterating over a list of items and you need to work on each of them only if some condition is verified. On the other hand, if you're iterating over a collection of items, and you have found one of them that satisfies some need you have, you may decide not to continue the loop entirely and therefore break out of it. There are countless possible scenarios, so it's better to see a couple of examples.
Let's say you want to apply a 20% discount to all products in a basket list for those which have an expiration date of today. The way you achieve this is to use the continue statement, which tells the looping construct (for
or while
) to immediately stop execution of the body and go to the next iteration, if any. This example will take us a little deeper down the rabbit whole, so be ready to jump.
discount.py
from datetime import date, timedelta today = date.today() tomorrow = today + timedelta(days=1) # today + 1 day is tomorrow products = [ {'sku': '1', 'expiration_date': today, 'price': 100.0}, {'sku': '2', 'expiration_date': tomorrow, 'price': 50}, {'sku': '3', 'expiration_date': today, 'price': 20}, ] for product in products: if product['expiration_date'] != today: continue product['price'] *= 0.8 # equivalent to applying 20% discount print( 'Price for sku', product['sku'], 'is now', product['price'])
You see we start by importing the date
and timedelta
objects, then we set up our products. Those with sku 1
and 3
have an expiration date of today
, which means we want to apply 20% discount on them. We loop over each product
and we inspect the expiration date. If it is not (inequality operator, !=
) today
, we don't want to execute the rest of the body suite, so we continue
.
Notice that is not important where in the body suite you place the continue
statement (you can even use it more than once). When you reach it, execution stops and goes back to the next iteration. If we run the discount.py
module, this is the output:
$ python discount.py Price for sku 1 is now 80.0 Price for sku 3 is now 16.0
Which shows you that the last two lines of the body haven't been executed for sku number 2.
Let's now see an example of breaking out of a loop. Say we want to tell if at least any of the elements in a list evaluates to True
when fed to the bool
function. Given that we need to know if there is at least one, when we find it we don't need to keep scanning the list any further. In Python code, this translates to using the break statement. Let's write this down into code:
any.py
items = [0, None, 0.0, True, 0, 7] # True and 7 evaluate to True found = False # this is called "flag" for item in items: print('scanning item', item) if item: found = True # we update the flag break if found: # we inspect the flag print('At least one item evaluates to True') else: print('All items evaluate to False')
The preceding code is such a common pattern in programming, you will see it a lot. When you inspect items this way, basically what you do is to set up a flag
variable, then start the inspection. If you find one element that matches your criteria (in this example, that evaluates to True
), then you update the flag and stop iterating. After iteration, you inspect the flag and take action accordingly. Execution yields:
$ python any.py scanning item 0 scanning item None scanning item 0.0 scanning item True At least one item evaluates to True
See how execution stopped after True
was found?
The break
statement acts exactly like the continue
one, in that it stops executing the body of the loop immediately, but also, prevents any other iteration to run, effectively breaking out of the loop.
The continue
and break
statements can be used together with no limitation in their number, both in the for
and while
looping constructs.
One of the features I've seen only in the Python language is the ability to have else
clauses after while
and for
loops. It's very rarely used, but it's definitely nice to have. In short, you can have an else
suite after a for
or while
loop. If the loop ends normally, because of exhaustion of the iterator (for
loop) or because the condition is finally not met (while
loop), then the else
suite (if present) is executed. In case execution is interrupted by a break
statement, the else
clause is not executed. Let's take an example of a for
loop that iterates over a group of items, looking for one that would match some condition. In case we don't find at least one that satisfies the condition, we want to raise an exception. This means we want to arrest the regular execution of the program and signal that there was an error, or exception, that we cannot deal with. Exceptions will be the subject of Chapter 7, Testing, Profiling, and Dealing with Exceptions, so don't worry if you don't fully understand them now. Just bear in mind that they will alter the regular flow of the code. Let me now show you two examples that do exactly the same thing, but one of them is using the special for
... else
syntax. Say that we want to find among a collection of people one that could drive a car.
for.no.else.py
class DriverException(Exception): pass people = [('James', 17), ('Kirk', 9), ('Lars', 13), ('Robert', 8)] driver = None for person, age in people: if age >= 18: driver = (person, age) break if driver is None: raise DriverException('Driver not found.')
Notice the flag
pattern again. We set driver to be None
, then if we find one we update the driver
flag, and then, at the end of the loop, we inspect it to see if one was found. I kind of have the feeling that those kids would drive a very metallic car, but anyway, notice that if a driver is not found, a DriverException
is raised, signaling the program that execution cannot continue (we're lacking the driver).
The same functionality can be rewritten a bit more elegantly using the following code:
for.else.py
class DriverException(Exception): pass people = [('James', 17), ('Kirk', 9), ('Lars', 13), ('Robert', 8)] for person, age in people: if age >= 18: driver = (person, age) break else: raise DriverException('Driver not found.')
Notice that we aren't forced to use the flag
pattern any more. The exception is raised as part of the for
loop logic, which makes good sense because the for
loop is checking on some condition. All we need is to set up a driver
object in case we find one, because the rest of the code is going to use that information somewhere. Notice the code is shorter and more elegant, because the logic is now correctly grouped together where it belongs.
Iterating over a sequence
Now we have all the tools to iterate over a sequence, so let's build on that example:
simple.for.2.py
surnames = ['Rivest', 'Shamir', 'Adleman'] for position in range(len(surnames)): print(position, surnames[position])
The preceding code adds a little bit of complexity to the game. Execution will show this result:
$ python simple.for.2.py 0 Rivest 1 Shamir 2 Adleman
Let's use the inside-out technique to break it down, ok? We start from the innermost part of what we're trying to understand, and we expand outwards. So, len(surnames)
is the length of the surnames
list: 3
. Therefore, range(len(surnames))
is actually transformed into range(3)
. This gives us the range [0, 3), which is basically a sequence (0, 1, 2)
. This means that the for
loop will run three iterations. In the first one, position
will take value 0
, while in the second one, it will take value 1
, and finally value 2
in the third and last iteration. What is (0, 1, 2)
, if not the possible indexing positions for the surnames
list? At position 0
we find 'Rivest'
, at position 1
, 'Shamir'
, and at position 2
, 'Adleman'
. If you are curious about what these three men created together, change print(position, surnames[position])
to print(surnames[position][0], end='')
add a final print()
outside of the loop, and run the code again.
Now, this style of looping is actually much closer to languages like Java or C++. In Python it's quite rare to see code like this. You can just iterate over any sequence or collection, so there is no need to get the list of positions and retrieve elements out of a sequence at each iteration. It's expensive, needlessly expensive. Let's change the example into a more Pythonic form:
simple.for.3.py
surnames = ['Rivest', 'Shamir', 'Adleman'] for surname in surnames: print(surname)
Now that's something! It's practically English. The for
loop can iterate over the surnames
list, and it gives back each element in order at each interaction. Running this code will print the three surnames, one at a time. It's much easier to read, right?
What if you wanted to print the position as well though? Or what if you actually needed it for any reason? Should you go back to the range(len(...))
form? No. You can use the enumerate
built-in function, like this:
simple.for.4.py
surnames = ['Rivest', 'Shamir', 'Adleman'] for position, surname in enumerate(surnames): print(position, surname)
This code is very interesting as well. Notice that enumerate gives back a 2-tuple (position, surname)
at each iteration, but still, it's much more readable (and more efficient) than the range(len(...))
example. You can call enumerate
with a start
parameter, like enumerate(iterable, start)
, and it will start from start
, rather than 0
. Just another little thing that shows you how much thought has been given in designing Python so that it makes your life easy.
Using a for
loop it is possible to iterate over lists, tuples, and in general anything that in Python is called iterable. This is a very important concept, so let's talk about it a bit more.
According to the Python documentation, an iterable is:
"An object capable of returning its members one at a time. Examples of iterables include all sequence types (such as
list
,str
, and tuple) and some non-sequence types likedict
,file
objects, and objects of any classes you define with an__iter__()
or__getitem__()
method. Iterables can be used in afor
loop and in many other places where a sequence is needed (zip()
,map()
, ...). When an iterable object is passed as an argument to the built-in functioniter()
, it returns an iterator for the object. This iterator is good for one pass over the set of values. When using iterables, it is usually not necessary to calliter()
or deal with iterator objects yourself. Thefor
statement does that automatically for you, creating a temporary unnamed variable to hold the iterator for the duration of the loop."
Simply put, what happens when you write for k in sequence: ... body ...
, is that the for
loop asks sequence
for the next element, it gets something back, it calls that something k
, and then executes its body. Then, once again, the for
loop asks sequence
again for the next element, it calls it k
again, and executes the body again, and so on and so forth, until the sequence is exhausted. Empty sequences will result in zero executions of the body.
Some data structures, when iterated over, produce their elements in order, like lists, tuples, and strings, while some others don't, like sets and dictionaries.
Python gives us the ability to iterate over iterables, using a type of object called iterator. According to the official documentation, an iterator is:
"An object representing a stream of data. Repeated calls to the iterator's
__next__()
method (or passing it to the built-in functionnext()
) return successive items in the stream. When no more data are available aStopIteration
exception is raised instead. At this point, the iterator object is exhausted and any further calls to its__next__()
method just raiseStopIteration
again. Iterators are required to have an__iter__()
method that returns the iterator object itself so every iterator is also iterable and may be used in most places where other iterables are accepted. One notable exception is code which attempts multiple iteration passes. A container object (such as alist
) produces a fresh new iterator each time you pass it to theiter()
function or use it in afor
loop. Attempting this with an iterator will just return the same exhausted iterator object used in the previous iteration pass, making it appear like an empty container."
Don't worry if you don't fully understand all the preceding legalese, you will in due time. I put it here as a handy reference for the future.
In practice, the whole iterable/iterator mechanism is somewhat hidden behind the code. Unless you need to code your own iterable or iterator for some reason, you won't have to worry about this too much. But it's very important to understand how Python handles this key aspect of control flow because it will shape the way you will write your code.
Let's see another example of how to iterate over two sequences of the same length, in order to work on their respective elements in pairs. Say we have a list of people and a list of numbers representing the age of the people in the first list. We want to print a pair person/age on one line for all of them. Let's start with an example and let's refine it gradually.
multiple.sequences.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] for position in range(len(people)): person = people[position] age = ages[position] print(person, age)
By now, this code should be pretty straightforward for you to understand. We need to iterate over the list of positions (0, 1, 2, 3) because we want to retrieve elements from two different lists. Executing it we get the following:
$ python multiple.sequences.py Jonas 25 Julio 30 Mike 31 Mez 39
This code is both inefficient and not Pythonic. Inefficient because retrieving an element given the position can be an expensive operation, and we're doing it from scratch at each iteration. The mail man doesn't go back to the beginning of the road each time he delivers a letter, right? He moves from house to house. From one to the next one. Let's try to make it better using enumerate:
multiple.sequences.enumerate.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] for position, person in enumerate(people): age = ages[position] print(person, age)
Better, but still not perfect. And still a bit ugly. We're iterating properly on people
, but we're still fetching age
using positional indexing, which we want to lose as well. Well, no worries, Python gives you the zip
function, remember? Let's use it!
multiple.sequences.zip.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] for person, age in zip(people, ages): print(person, age)
Ah! So much better! Once again, compare the preceding code with the first example and admire Python's elegance. The reason I wanted to show this example is twofold. On the one hand, I wanted to give you an idea of how shorter the code in Python can be compared to other languages where the syntax doesn't allow you to iterate over sequences or collections as easily. And on the other hand, and much more importantly, notice that when the for
loop asks zip(sequenceA, sequenceB)
for the next element, it gets back a tuple
, not just a single object. It gets back a tuple
with as many elements as the number of sequences we feed to the zip
function. Let's expand a little on the previous example in two ways: using explicit and implicit assignment:
multiple.sequences.explicit.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] nationalities = ['Belgium', 'Spain', 'England', 'Bangladesh'] for person, age, nationality in zip(people, ages, nationalities): print(person, age, nationality)
In the preceding code, we added the nationalities list. Now that we feed three sequences to the zip
function, the for loop gets back a 3-tuple at each iteration. Notice that the position of the elements in the tuple respects the position of the sequences in the zip
call. Executing the code will yield the following result:
$ python multiple.sequences.explicit.py Jonas 25 Belgium Julio 30 Spain Mike 31 England Mez 39 Bangladesh
Sometimes, for reasons that may not be clear in a simple example like the preceding one, you may want to explode the tuple within the body of the for
loop. If that is your desire, it's perfectly possible to do so.
multiple.sequences.implicit.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] nationalities = ['Belgium', 'Spain', 'England', 'Bangladesh'] for data in zip(people, ages, nationalities): person, age, nationality = data print(person, age, nationality)
It's basically doing what the for
loop does automatically for you, but in some cases you may want to do it yourself. Here, the 3-tuple data
that comes from zip(...)
, is exploded within the body of the for
loop into three variables: person
, age
, and nationality
.
In the preceding pages, we saw the for
loop in action. It's incredibly useful when you need to loop over a sequence or a collection. The key point to keep in mind, when you need to be able to discriminate which looping construct to use, is that the for
loop rocks when you have to iterate over a finite amount of elements. It can be a huge amount, but still, something that at some point ends.
There are other cases though, when you just need to loop until some condition is satisfied, or even loop indefinitely until the application is stopped. Cases where we don't really have something to iterate on, and therefore the for
loop would be a poor choice. But fear not, for these cases Python provides us with the while
loop.
The while
loop is similar to the for
loop, in that they both loop and at each iteration they execute a body of instructions. What is different between them is that the while
loop doesn't loop over a sequence (it can, but you have to manually write the logic and it wouldn't make any sense, you would just want to use a for
loop), rather, it loops as long as a certain condition is satisfied. When the condition is no longer satisfied, the loop ends.
As usual, let's see an example which will clarify everything for us. We want to print the binary representation of a positive number. In order to do so, we repeatedly divide the number by two, collecting the remainder, and then produce the inverse of the list of remainders. Let me give you a small example using number 6, which is 110 in binary.
6 / 2 = 3 (remainder: 0) 3 / 2 = 1 (remainder: 1) 1 / 2 = 0 (remainder: 1) List of remainders: 0, 1, 1. Inverse is 1, 1, 0, which is also the binary representation of 6: 110
Let's write some code to calculate the binary representation for number 39: 1001112.
binary.py
n = 39 remainders = [] while n > 0: remainder = n % 2 # remainder of division by 2 remainders.append(remainder) # we keep track of remainders n //= 2 # we divide n by 2 # reassign the list to its reversed copy and print it remainders = remainders[::-1] print(remainders)
In the preceding code, I highlighted two things: n > 0
, which is the condition to keep looping, and remainders[::-1]
which is a nice and easy way to get the reversed version of a list (missing start
and end
parameters, step = -1
, produces the same list, from end
to start
, in reverse order). We can make the code a little shorter (and more Pythonic), by using the divmod
function, which is called with a number and a divisor, and returns a tuple with the result of the integer division and its remainder. For example, divmod(13, 5)
would return (2, 3)
, and indeed 5 * 2 + 3 = 13.
binary.2.py
n = 39
remainders = []
while n > 0:
n, remainder = divmod(n, 2)
remainders.append(remainder)
# reassign the list to its reversed copy and print it
remainders = remainders[::-1]
print(remainders)
In the preceding code, we have reassigned n to the result of the division by 2, and the remainder, in one single line.
Notice that the condition in a while
loop is a condition to continue looping. If it evaluates to True
, then the body is executed and then another evaluation follows, and so on, until the condition evaluates to False
. When that happens, the loop is exited immediately without executing its body.
Note
If the condition never evaluates to False
, the loop becomes a so called infinite loop. Infinite loops are used for example when polling from network devices: you ask the socket if there is any data, you do something with it if there is any, then you sleep for a small amount of time, and then you ask the socket again, over and over again, without ever stopping.
Having the ability to loop over a condition, or to loop indefinitely, is the reason why the for
loop alone is not enough, and therefore Python provides the while
loop.
Just for fun, let's adapt one of the examples (multiple.sequences.py
) using the while logic.
multiple.sequences.while.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] position = 0 while position < len(people): person = people[position] age = ages[position] print(person, age) position += 1
In the preceding code, I have highlighted the initialization, condition, and update of the variable position
, which makes it possible to simulate the equivalent for
loop code by handling the iteration variable manually. Everything that can be done with a for
loop can also be done with a while
loop, even though you can see there's a bit of boilerplate you have to go through in order to achieve the same result. The opposite is also true, but simulating a never ending while
loop using a for
loop requires some real trickery, so why would you do that? Use the right tool for the job, and 99.9% of the times you'll be fine.
So, to recap, use a for
loop when you need to iterate over one (or a combination of) iterable, and a while
loop when you need to loop according to a condition being satisfied or not. If you keep in mind the difference between the two purposes, you will never choose the wrong looping construct.
Let's now see how to alter the normal flow of a loop.
According to the task at hand, sometimes you will need to alter the regular flow of a loop. You can either skip a single iteration (as many times you want), or you can break out of the loop entirely. A common use case for skipping iterations is for example when you're iterating over a list of items and you need to work on each of them only if some condition is verified. On the other hand, if you're iterating over a collection of items, and you have found one of them that satisfies some need you have, you may decide not to continue the loop entirely and therefore break out of it. There are countless possible scenarios, so it's better to see a couple of examples.
Let's say you want to apply a 20% discount to all products in a basket list for those which have an expiration date of today. The way you achieve this is to use the continue statement, which tells the looping construct (for
or while
) to immediately stop execution of the body and go to the next iteration, if any. This example will take us a little deeper down the rabbit whole, so be ready to jump.
discount.py
from datetime import date, timedelta today = date.today() tomorrow = today + timedelta(days=1) # today + 1 day is tomorrow products = [ {'sku': '1', 'expiration_date': today, 'price': 100.0}, {'sku': '2', 'expiration_date': tomorrow, 'price': 50}, {'sku': '3', 'expiration_date': today, 'price': 20}, ] for product in products: if product['expiration_date'] != today: continue product['price'] *= 0.8 # equivalent to applying 20% discount print( 'Price for sku', product['sku'], 'is now', product['price'])
You see we start by importing the date
and timedelta
objects, then we set up our products. Those with sku 1
and 3
have an expiration date of today
, which means we want to apply 20% discount on them. We loop over each product
and we inspect the expiration date. If it is not (inequality operator, !=
) today
, we don't want to execute the rest of the body suite, so we continue
.
Notice that is not important where in the body suite you place the continue
statement (you can even use it more than once). When you reach it, execution stops and goes back to the next iteration. If we run the discount.py
module, this is the output:
$ python discount.py Price for sku 1 is now 80.0 Price for sku 3 is now 16.0
Which shows you that the last two lines of the body haven't been executed for sku number 2.
Let's now see an example of breaking out of a loop. Say we want to tell if at least any of the elements in a list evaluates to True
when fed to the bool
function. Given that we need to know if there is at least one, when we find it we don't need to keep scanning the list any further. In Python code, this translates to using the break statement. Let's write this down into code:
any.py
items = [0, None, 0.0, True, 0, 7] # True and 7 evaluate to True found = False # this is called "flag" for item in items: print('scanning item', item) if item: found = True # we update the flag break if found: # we inspect the flag print('At least one item evaluates to True') else: print('All items evaluate to False')
The preceding code is such a common pattern in programming, you will see it a lot. When you inspect items this way, basically what you do is to set up a flag
variable, then start the inspection. If you find one element that matches your criteria (in this example, that evaluates to True
), then you update the flag and stop iterating. After iteration, you inspect the flag and take action accordingly. Execution yields:
$ python any.py scanning item 0 scanning item None scanning item 0.0 scanning item True At least one item evaluates to True
See how execution stopped after True
was found?
The break
statement acts exactly like the continue
one, in that it stops executing the body of the loop immediately, but also, prevents any other iteration to run, effectively breaking out of the loop.
The continue
and break
statements can be used together with no limitation in their number, both in the for
and while
looping constructs.
One of the features I've seen only in the Python language is the ability to have else
clauses after while
and for
loops. It's very rarely used, but it's definitely nice to have. In short, you can have an else
suite after a for
or while
loop. If the loop ends normally, because of exhaustion of the iterator (for
loop) or because the condition is finally not met (while
loop), then the else
suite (if present) is executed. In case execution is interrupted by a break
statement, the else
clause is not executed. Let's take an example of a for
loop that iterates over a group of items, looking for one that would match some condition. In case we don't find at least one that satisfies the condition, we want to raise an exception. This means we want to arrest the regular execution of the program and signal that there was an error, or exception, that we cannot deal with. Exceptions will be the subject of Chapter 7, Testing, Profiling, and Dealing with Exceptions, so don't worry if you don't fully understand them now. Just bear in mind that they will alter the regular flow of the code. Let me now show you two examples that do exactly the same thing, but one of them is using the special for
... else
syntax. Say that we want to find among a collection of people one that could drive a car.
for.no.else.py
class DriverException(Exception): pass people = [('James', 17), ('Kirk', 9), ('Lars', 13), ('Robert', 8)] driver = None for person, age in people: if age >= 18: driver = (person, age) break if driver is None: raise DriverException('Driver not found.')
Notice the flag
pattern again. We set driver to be None
, then if we find one we update the driver
flag, and then, at the end of the loop, we inspect it to see if one was found. I kind of have the feeling that those kids would drive a very metallic car, but anyway, notice that if a driver is not found, a DriverException
is raised, signaling the program that execution cannot continue (we're lacking the driver).
The same functionality can be rewritten a bit more elegantly using the following code:
for.else.py
class DriverException(Exception): pass people = [('James', 17), ('Kirk', 9), ('Lars', 13), ('Robert', 8)] for person, age in people: if age >= 18: driver = (person, age) break else: raise DriverException('Driver not found.')
Notice that we aren't forced to use the flag
pattern any more. The exception is raised as part of the for
loop logic, which makes good sense because the for
loop is checking on some condition. All we need is to set up a driver
object in case we find one, because the rest of the code is going to use that information somewhere. Notice the code is shorter and more elegant, because the logic is now correctly grouped together where it belongs.
Iterators and iterables
According to the Python documentation, an iterable is:
"An object capable of returning its members one at a time. Examples of iterables include all sequence types (such as
list
,str
, and tuple) and some non-sequence types likedict
,file
objects, and objects of any classes you define with an__iter__()
or__getitem__()
method. Iterables can be used in afor
loop and in many other places where a sequence is needed (zip()
,map()
, ...). When an iterable object is passed as an argument to the built-in functioniter()
, it returns an iterator for the object. This iterator is good for one pass over the set of values. When using iterables, it is usually not necessary to calliter()
or deal with iterator objects yourself. Thefor
statement does that automatically for you, creating a temporary unnamed variable to hold the iterator for the duration of the loop."
Simply put, what happens when you write for k in sequence: ... body ...
, is that the for
loop asks sequence
for the next element, it gets something back, it calls that something k
, and then executes its body. Then, once again, the for
loop asks sequence
again for the next element, it calls it k
again, and executes the body again, and so on and so forth, until the sequence is exhausted. Empty sequences will result in zero executions of the body.
Some data structures, when iterated over, produce their elements in order, like lists, tuples, and strings, while some others don't, like sets and dictionaries.
Python gives us the ability to iterate over iterables, using a type of object called iterator. According to the official documentation, an iterator is:
"An object representing a stream of data. Repeated calls to the iterator's
__next__()
method (or passing it to the built-in functionnext()
) return successive items in the stream. When no more data are available aStopIteration
exception is raised instead. At this point, the iterator object is exhausted and any further calls to its__next__()
method just raiseStopIteration
again. Iterators are required to have an__iter__()
method that returns the iterator object itself so every iterator is also iterable and may be used in most places where other iterables are accepted. One notable exception is code which attempts multiple iteration passes. A container object (such as alist
) produces a fresh new iterator each time you pass it to theiter()
function or use it in afor
loop. Attempting this with an iterator will just return the same exhausted iterator object used in the previous iteration pass, making it appear like an empty container."
Don't worry if you don't fully understand all the preceding legalese, you will in due time. I put it here as a handy reference for the future.
In practice, the whole iterable/iterator mechanism is somewhat hidden behind the code. Unless you need to code your own iterable or iterator for some reason, you won't have to worry about this too much. But it's very important to understand how Python handles this key aspect of control flow because it will shape the way you will write your code.
Let's see another example of how to iterate over two sequences of the same length, in order to work on their respective elements in pairs. Say we have a list of people and a list of numbers representing the age of the people in the first list. We want to print a pair person/age on one line for all of them. Let's start with an example and let's refine it gradually.
multiple.sequences.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] for position in range(len(people)): person = people[position] age = ages[position] print(person, age)
By now, this code should be pretty straightforward for you to understand. We need to iterate over the list of positions (0, 1, 2, 3) because we want to retrieve elements from two different lists. Executing it we get the following:
$ python multiple.sequences.py Jonas 25 Julio 30 Mike 31 Mez 39
This code is both inefficient and not Pythonic. Inefficient because retrieving an element given the position can be an expensive operation, and we're doing it from scratch at each iteration. The mail man doesn't go back to the beginning of the road each time he delivers a letter, right? He moves from house to house. From one to the next one. Let's try to make it better using enumerate:
multiple.sequences.enumerate.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] for position, person in enumerate(people): age = ages[position] print(person, age)
Better, but still not perfect. And still a bit ugly. We're iterating properly on people
, but we're still fetching age
using positional indexing, which we want to lose as well. Well, no worries, Python gives you the zip
function, remember? Let's use it!
multiple.sequences.zip.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] for person, age in zip(people, ages): print(person, age)
Ah! So much better! Once again, compare the preceding code with the first example and admire Python's elegance. The reason I wanted to show this example is twofold. On the one hand, I wanted to give you an idea of how shorter the code in Python can be compared to other languages where the syntax doesn't allow you to iterate over sequences or collections as easily. And on the other hand, and much more importantly, notice that when the for
loop asks zip(sequenceA, sequenceB)
for the next element, it gets back a tuple
, not just a single object. It gets back a tuple
with as many elements as the number of sequences we feed to the zip
function. Let's expand a little on the previous example in two ways: using explicit and implicit assignment:
multiple.sequences.explicit.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] nationalities = ['Belgium', 'Spain', 'England', 'Bangladesh'] for person, age, nationality in zip(people, ages, nationalities): print(person, age, nationality)
In the preceding code, we added the nationalities list. Now that we feed three sequences to the zip
function, the for loop gets back a 3-tuple at each iteration. Notice that the position of the elements in the tuple respects the position of the sequences in the zip
call. Executing the code will yield the following result:
$ python multiple.sequences.explicit.py Jonas 25 Belgium Julio 30 Spain Mike 31 England Mez 39 Bangladesh
Sometimes, for reasons that may not be clear in a simple example like the preceding one, you may want to explode the tuple within the body of the for
loop. If that is your desire, it's perfectly possible to do so.
multiple.sequences.implicit.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] nationalities = ['Belgium', 'Spain', 'England', 'Bangladesh'] for data in zip(people, ages, nationalities): person, age, nationality = data print(person, age, nationality)
It's basically doing what the for
loop does automatically for you, but in some cases you may want to do it yourself. Here, the 3-tuple data
that comes from zip(...)
, is exploded within the body of the for
loop into three variables: person
, age
, and nationality
.
In the preceding pages, we saw the for
loop in action. It's incredibly useful when you need to loop over a sequence or a collection. The key point to keep in mind, when you need to be able to discriminate which looping construct to use, is that the for
loop rocks when you have to iterate over a finite amount of elements. It can be a huge amount, but still, something that at some point ends.
There are other cases though, when you just need to loop until some condition is satisfied, or even loop indefinitely until the application is stopped. Cases where we don't really have something to iterate on, and therefore the for
loop would be a poor choice. But fear not, for these cases Python provides us with the while
loop.
The while
loop is similar to the for
loop, in that they both loop and at each iteration they execute a body of instructions. What is different between them is that the while
loop doesn't loop over a sequence (it can, but you have to manually write the logic and it wouldn't make any sense, you would just want to use a for
loop), rather, it loops as long as a certain condition is satisfied. When the condition is no longer satisfied, the loop ends.
As usual, let's see an example which will clarify everything for us. We want to print the binary representation of a positive number. In order to do so, we repeatedly divide the number by two, collecting the remainder, and then produce the inverse of the list of remainders. Let me give you a small example using number 6, which is 110 in binary.
6 / 2 = 3 (remainder: 0) 3 / 2 = 1 (remainder: 1) 1 / 2 = 0 (remainder: 1) List of remainders: 0, 1, 1. Inverse is 1, 1, 0, which is also the binary representation of 6: 110
Let's write some code to calculate the binary representation for number 39: 1001112.
binary.py
n = 39 remainders = [] while n > 0: remainder = n % 2 # remainder of division by 2 remainders.append(remainder) # we keep track of remainders n //= 2 # we divide n by 2 # reassign the list to its reversed copy and print it remainders = remainders[::-1] print(remainders)
In the preceding code, I highlighted two things: n > 0
, which is the condition to keep looping, and remainders[::-1]
which is a nice and easy way to get the reversed version of a list (missing start
and end
parameters, step = -1
, produces the same list, from end
to start
, in reverse order). We can make the code a little shorter (and more Pythonic), by using the divmod
function, which is called with a number and a divisor, and returns a tuple with the result of the integer division and its remainder. For example, divmod(13, 5)
would return (2, 3)
, and indeed 5 * 2 + 3 = 13.
binary.2.py
n = 39
remainders = []
while n > 0:
n, remainder = divmod(n, 2)
remainders.append(remainder)
# reassign the list to its reversed copy and print it
remainders = remainders[::-1]
print(remainders)
In the preceding code, we have reassigned n to the result of the division by 2, and the remainder, in one single line.
Notice that the condition in a while
loop is a condition to continue looping. If it evaluates to True
, then the body is executed and then another evaluation follows, and so on, until the condition evaluates to False
. When that happens, the loop is exited immediately without executing its body.
Note
If the condition never evaluates to False
, the loop becomes a so called infinite loop. Infinite loops are used for example when polling from network devices: you ask the socket if there is any data, you do something with it if there is any, then you sleep for a small amount of time, and then you ask the socket again, over and over again, without ever stopping.
Having the ability to loop over a condition, or to loop indefinitely, is the reason why the for
loop alone is not enough, and therefore Python provides the while
loop.
Just for fun, let's adapt one of the examples (multiple.sequences.py
) using the while logic.
multiple.sequences.while.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] position = 0 while position < len(people): person = people[position] age = ages[position] print(person, age) position += 1
In the preceding code, I have highlighted the initialization, condition, and update of the variable position
, which makes it possible to simulate the equivalent for
loop code by handling the iteration variable manually. Everything that can be done with a for
loop can also be done with a while
loop, even though you can see there's a bit of boilerplate you have to go through in order to achieve the same result. The opposite is also true, but simulating a never ending while
loop using a for
loop requires some real trickery, so why would you do that? Use the right tool for the job, and 99.9% of the times you'll be fine.
So, to recap, use a for
loop when you need to iterate over one (or a combination of) iterable, and a while
loop when you need to loop according to a condition being satisfied or not. If you keep in mind the difference between the two purposes, you will never choose the wrong looping construct.
Let's now see how to alter the normal flow of a loop.
According to the task at hand, sometimes you will need to alter the regular flow of a loop. You can either skip a single iteration (as many times you want), or you can break out of the loop entirely. A common use case for skipping iterations is for example when you're iterating over a list of items and you need to work on each of them only if some condition is verified. On the other hand, if you're iterating over a collection of items, and you have found one of them that satisfies some need you have, you may decide not to continue the loop entirely and therefore break out of it. There are countless possible scenarios, so it's better to see a couple of examples.
Let's say you want to apply a 20% discount to all products in a basket list for those which have an expiration date of today. The way you achieve this is to use the continue statement, which tells the looping construct (for
or while
) to immediately stop execution of the body and go to the next iteration, if any. This example will take us a little deeper down the rabbit whole, so be ready to jump.
discount.py
from datetime import date, timedelta today = date.today() tomorrow = today + timedelta(days=1) # today + 1 day is tomorrow products = [ {'sku': '1', 'expiration_date': today, 'price': 100.0}, {'sku': '2', 'expiration_date': tomorrow, 'price': 50}, {'sku': '3', 'expiration_date': today, 'price': 20}, ] for product in products: if product['expiration_date'] != today: continue product['price'] *= 0.8 # equivalent to applying 20% discount print( 'Price for sku', product['sku'], 'is now', product['price'])
You see we start by importing the date
and timedelta
objects, then we set up our products. Those with sku 1
and 3
have an expiration date of today
, which means we want to apply 20% discount on them. We loop over each product
and we inspect the expiration date. If it is not (inequality operator, !=
) today
, we don't want to execute the rest of the body suite, so we continue
.
Notice that is not important where in the body suite you place the continue
statement (you can even use it more than once). When you reach it, execution stops and goes back to the next iteration. If we run the discount.py
module, this is the output:
$ python discount.py Price for sku 1 is now 80.0 Price for sku 3 is now 16.0
Which shows you that the last two lines of the body haven't been executed for sku number 2.
Let's now see an example of breaking out of a loop. Say we want to tell if at least any of the elements in a list evaluates to True
when fed to the bool
function. Given that we need to know if there is at least one, when we find it we don't need to keep scanning the list any further. In Python code, this translates to using the break statement. Let's write this down into code:
any.py
items = [0, None, 0.0, True, 0, 7] # True and 7 evaluate to True found = False # this is called "flag" for item in items: print('scanning item', item) if item: found = True # we update the flag break if found: # we inspect the flag print('At least one item evaluates to True') else: print('All items evaluate to False')
The preceding code is such a common pattern in programming, you will see it a lot. When you inspect items this way, basically what you do is to set up a flag
variable, then start the inspection. If you find one element that matches your criteria (in this example, that evaluates to True
), then you update the flag and stop iterating. After iteration, you inspect the flag and take action accordingly. Execution yields:
$ python any.py scanning item 0 scanning item None scanning item 0.0 scanning item True At least one item evaluates to True
See how execution stopped after True
was found?
The break
statement acts exactly like the continue
one, in that it stops executing the body of the loop immediately, but also, prevents any other iteration to run, effectively breaking out of the loop.
The continue
and break
statements can be used together with no limitation in their number, both in the for
and while
looping constructs.
One of the features I've seen only in the Python language is the ability to have else
clauses after while
and for
loops. It's very rarely used, but it's definitely nice to have. In short, you can have an else
suite after a for
or while
loop. If the loop ends normally, because of exhaustion of the iterator (for
loop) or because the condition is finally not met (while
loop), then the else
suite (if present) is executed. In case execution is interrupted by a break
statement, the else
clause is not executed. Let's take an example of a for
loop that iterates over a group of items, looking for one that would match some condition. In case we don't find at least one that satisfies the condition, we want to raise an exception. This means we want to arrest the regular execution of the program and signal that there was an error, or exception, that we cannot deal with. Exceptions will be the subject of Chapter 7, Testing, Profiling, and Dealing with Exceptions, so don't worry if you don't fully understand them now. Just bear in mind that they will alter the regular flow of the code. Let me now show you two examples that do exactly the same thing, but one of them is using the special for
... else
syntax. Say that we want to find among a collection of people one that could drive a car.
for.no.else.py
class DriverException(Exception): pass people = [('James', 17), ('Kirk', 9), ('Lars', 13), ('Robert', 8)] driver = None for person, age in people: if age >= 18: driver = (person, age) break if driver is None: raise DriverException('Driver not found.')
Notice the flag
pattern again. We set driver to be None
, then if we find one we update the driver
flag, and then, at the end of the loop, we inspect it to see if one was found. I kind of have the feeling that those kids would drive a very metallic car, but anyway, notice that if a driver is not found, a DriverException
is raised, signaling the program that execution cannot continue (we're lacking the driver).
The same functionality can be rewritten a bit more elegantly using the following code:
for.else.py
class DriverException(Exception): pass people = [('James', 17), ('Kirk', 9), ('Lars', 13), ('Robert', 8)] for person, age in people: if age >= 18: driver = (person, age) break else: raise DriverException('Driver not found.')
Notice that we aren't forced to use the flag
pattern any more. The exception is raised as part of the for
loop logic, which makes good sense because the for
loop is checking on some condition. All we need is to set up a driver
object in case we find one, because the rest of the code is going to use that information somewhere. Notice the code is shorter and more elegant, because the logic is now correctly grouped together where it belongs.
Iterating over multiple sequences
Let's see another example of how to iterate over two sequences of the same length, in order to work on their respective elements in pairs. Say we have a list of people and a list of numbers representing the age of the people in the first list. We want to print a pair person/age on one line for all of them. Let's start with an example and let's refine it gradually.
multiple.sequences.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] for position in range(len(people)): person = people[position] age = ages[position] print(person, age)
By now, this code should be pretty straightforward for you to understand. We need to iterate over the list of positions (0, 1, 2, 3) because we want to retrieve elements from two different lists. Executing it we get the following:
$ python multiple.sequences.py Jonas 25 Julio 30 Mike 31 Mez 39
This code is both inefficient and not Pythonic. Inefficient because retrieving an element given the position can be an expensive operation, and we're doing it from scratch at each iteration. The mail man doesn't go back to the beginning of the road each time he delivers a letter, right? He moves from house to house. From one to the next one. Let's try to make it better using enumerate:
multiple.sequences.enumerate.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] for position, person in enumerate(people): age = ages[position] print(person, age)
Better, but still not perfect. And still a bit ugly. We're iterating properly on people
, but we're still fetching age
using positional indexing, which we want to lose as well. Well, no worries, Python gives you the zip
function, remember? Let's use it!
multiple.sequences.zip.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] for person, age in zip(people, ages): print(person, age)
Ah! So much better! Once again, compare the preceding code with the first example and admire Python's elegance. The reason I wanted to show this example is twofold. On the one hand, I wanted to give you an idea of how shorter the code in Python can be compared to other languages where the syntax doesn't allow you to iterate over sequences or collections as easily. And on the other hand, and much more importantly, notice that when the for
loop asks zip(sequenceA, sequenceB)
for the next element, it gets back a tuple
, not just a single object. It gets back a tuple
with as many elements as the number of sequences we feed to the zip
function. Let's expand a little on the previous example in two ways: using explicit and implicit assignment:
multiple.sequences.explicit.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] nationalities = ['Belgium', 'Spain', 'England', 'Bangladesh'] for person, age, nationality in zip(people, ages, nationalities): print(person, age, nationality)
In the preceding code, we added the nationalities list. Now that we feed three sequences to the zip
function, the for loop gets back a 3-tuple at each iteration. Notice that the position of the elements in the tuple respects the position of the sequences in the zip
call. Executing the code will yield the following result:
$ python multiple.sequences.explicit.py Jonas 25 Belgium Julio 30 Spain Mike 31 England Mez 39 Bangladesh
Sometimes, for reasons that may not be clear in a simple example like the preceding one, you may want to explode the tuple within the body of the for
loop. If that is your desire, it's perfectly possible to do so.
multiple.sequences.implicit.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] nationalities = ['Belgium', 'Spain', 'England', 'Bangladesh'] for data in zip(people, ages, nationalities): person, age, nationality = data print(person, age, nationality)
It's basically doing what the for
loop does automatically for you, but in some cases you may want to do it yourself. Here, the 3-tuple data
that comes from zip(...)
, is exploded within the body of the for
loop into three variables: person
, age
, and nationality
.
In the preceding pages, we saw the for
loop in action. It's incredibly useful when you need to loop over a sequence or a collection. The key point to keep in mind, when you need to be able to discriminate which looping construct to use, is that the for
loop rocks when you have to iterate over a finite amount of elements. It can be a huge amount, but still, something that at some point ends.
There are other cases though, when you just need to loop until some condition is satisfied, or even loop indefinitely until the application is stopped. Cases where we don't really have something to iterate on, and therefore the for
loop would be a poor choice. But fear not, for these cases Python provides us with the while
loop.
The while
loop is similar to the for
loop, in that they both loop and at each iteration they execute a body of instructions. What is different between them is that the while
loop doesn't loop over a sequence (it can, but you have to manually write the logic and it wouldn't make any sense, you would just want to use a for
loop), rather, it loops as long as a certain condition is satisfied. When the condition is no longer satisfied, the loop ends.
As usual, let's see an example which will clarify everything for us. We want to print the binary representation of a positive number. In order to do so, we repeatedly divide the number by two, collecting the remainder, and then produce the inverse of the list of remainders. Let me give you a small example using number 6, which is 110 in binary.
6 / 2 = 3 (remainder: 0) 3 / 2 = 1 (remainder: 1) 1 / 2 = 0 (remainder: 1) List of remainders: 0, 1, 1. Inverse is 1, 1, 0, which is also the binary representation of 6: 110
Let's write some code to calculate the binary representation for number 39: 1001112.
binary.py
n = 39 remainders = [] while n > 0: remainder = n % 2 # remainder of division by 2 remainders.append(remainder) # we keep track of remainders n //= 2 # we divide n by 2 # reassign the list to its reversed copy and print it remainders = remainders[::-1] print(remainders)
In the preceding code, I highlighted two things: n > 0
, which is the condition to keep looping, and remainders[::-1]
which is a nice and easy way to get the reversed version of a list (missing start
and end
parameters, step = -1
, produces the same list, from end
to start
, in reverse order). We can make the code a little shorter (and more Pythonic), by using the divmod
function, which is called with a number and a divisor, and returns a tuple with the result of the integer division and its remainder. For example, divmod(13, 5)
would return (2, 3)
, and indeed 5 * 2 + 3 = 13.
binary.2.py
n = 39
remainders = []
while n > 0:
n, remainder = divmod(n, 2)
remainders.append(remainder)
# reassign the list to its reversed copy and print it
remainders = remainders[::-1]
print(remainders)
In the preceding code, we have reassigned n to the result of the division by 2, and the remainder, in one single line.
Notice that the condition in a while
loop is a condition to continue looping. If it evaluates to True
, then the body is executed and then another evaluation follows, and so on, until the condition evaluates to False
. When that happens, the loop is exited immediately without executing its body.
Note
If the condition never evaluates to False
, the loop becomes a so called infinite loop. Infinite loops are used for example when polling from network devices: you ask the socket if there is any data, you do something with it if there is any, then you sleep for a small amount of time, and then you ask the socket again, over and over again, without ever stopping.
Having the ability to loop over a condition, or to loop indefinitely, is the reason why the for
loop alone is not enough, and therefore Python provides the while
loop.
Just for fun, let's adapt one of the examples (multiple.sequences.py
) using the while logic.
multiple.sequences.while.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] position = 0 while position < len(people): person = people[position] age = ages[position] print(person, age) position += 1
In the preceding code, I have highlighted the initialization, condition, and update of the variable position
, which makes it possible to simulate the equivalent for
loop code by handling the iteration variable manually. Everything that can be done with a for
loop can also be done with a while
loop, even though you can see there's a bit of boilerplate you have to go through in order to achieve the same result. The opposite is also true, but simulating a never ending while
loop using a for
loop requires some real trickery, so why would you do that? Use the right tool for the job, and 99.9% of the times you'll be fine.
So, to recap, use a for
loop when you need to iterate over one (or a combination of) iterable, and a while
loop when you need to loop according to a condition being satisfied or not. If you keep in mind the difference between the two purposes, you will never choose the wrong looping construct.
Let's now see how to alter the normal flow of a loop.
According to the task at hand, sometimes you will need to alter the regular flow of a loop. You can either skip a single iteration (as many times you want), or you can break out of the loop entirely. A common use case for skipping iterations is for example when you're iterating over a list of items and you need to work on each of them only if some condition is verified. On the other hand, if you're iterating over a collection of items, and you have found one of them that satisfies some need you have, you may decide not to continue the loop entirely and therefore break out of it. There are countless possible scenarios, so it's better to see a couple of examples.
Let's say you want to apply a 20% discount to all products in a basket list for those which have an expiration date of today. The way you achieve this is to use the continue statement, which tells the looping construct (for
or while
) to immediately stop execution of the body and go to the next iteration, if any. This example will take us a little deeper down the rabbit whole, so be ready to jump.
discount.py
from datetime import date, timedelta today = date.today() tomorrow = today + timedelta(days=1) # today + 1 day is tomorrow products = [ {'sku': '1', 'expiration_date': today, 'price': 100.0}, {'sku': '2', 'expiration_date': tomorrow, 'price': 50}, {'sku': '3', 'expiration_date': today, 'price': 20}, ] for product in products: if product['expiration_date'] != today: continue product['price'] *= 0.8 # equivalent to applying 20% discount print( 'Price for sku', product['sku'], 'is now', product['price'])
You see we start by importing the date
and timedelta
objects, then we set up our products. Those with sku 1
and 3
have an expiration date of today
, which means we want to apply 20% discount on them. We loop over each product
and we inspect the expiration date. If it is not (inequality operator, !=
) today
, we don't want to execute the rest of the body suite, so we continue
.
Notice that is not important where in the body suite you place the continue
statement (you can even use it more than once). When you reach it, execution stops and goes back to the next iteration. If we run the discount.py
module, this is the output:
$ python discount.py Price for sku 1 is now 80.0 Price for sku 3 is now 16.0
Which shows you that the last two lines of the body haven't been executed for sku number 2.
Let's now see an example of breaking out of a loop. Say we want to tell if at least any of the elements in a list evaluates to True
when fed to the bool
function. Given that we need to know if there is at least one, when we find it we don't need to keep scanning the list any further. In Python code, this translates to using the break statement. Let's write this down into code:
any.py
items = [0, None, 0.0, True, 0, 7] # True and 7 evaluate to True found = False # this is called "flag" for item in items: print('scanning item', item) if item: found = True # we update the flag break if found: # we inspect the flag print('At least one item evaluates to True') else: print('All items evaluate to False')
The preceding code is such a common pattern in programming, you will see it a lot. When you inspect items this way, basically what you do is to set up a flag
variable, then start the inspection. If you find one element that matches your criteria (in this example, that evaluates to True
), then you update the flag and stop iterating. After iteration, you inspect the flag and take action accordingly. Execution yields:
$ python any.py scanning item 0 scanning item None scanning item 0.0 scanning item True At least one item evaluates to True
See how execution stopped after True
was found?
The break
statement acts exactly like the continue
one, in that it stops executing the body of the loop immediately, but also, prevents any other iteration to run, effectively breaking out of the loop.
The continue
and break
statements can be used together with no limitation in their number, both in the for
and while
looping constructs.
One of the features I've seen only in the Python language is the ability to have else
clauses after while
and for
loops. It's very rarely used, but it's definitely nice to have. In short, you can have an else
suite after a for
or while
loop. If the loop ends normally, because of exhaustion of the iterator (for
loop) or because the condition is finally not met (while
loop), then the else
suite (if present) is executed. In case execution is interrupted by a break
statement, the else
clause is not executed. Let's take an example of a for
loop that iterates over a group of items, looking for one that would match some condition. In case we don't find at least one that satisfies the condition, we want to raise an exception. This means we want to arrest the regular execution of the program and signal that there was an error, or exception, that we cannot deal with. Exceptions will be the subject of Chapter 7, Testing, Profiling, and Dealing with Exceptions, so don't worry if you don't fully understand them now. Just bear in mind that they will alter the regular flow of the code. Let me now show you two examples that do exactly the same thing, but one of them is using the special for
... else
syntax. Say that we want to find among a collection of people one that could drive a car.
for.no.else.py
class DriverException(Exception): pass people = [('James', 17), ('Kirk', 9), ('Lars', 13), ('Robert', 8)] driver = None for person, age in people: if age >= 18: driver = (person, age) break if driver is None: raise DriverException('Driver not found.')
Notice the flag
pattern again. We set driver to be None
, then if we find one we update the driver
flag, and then, at the end of the loop, we inspect it to see if one was found. I kind of have the feeling that those kids would drive a very metallic car, but anyway, notice that if a driver is not found, a DriverException
is raised, signaling the program that execution cannot continue (we're lacking the driver).
The same functionality can be rewritten a bit more elegantly using the following code:
for.else.py
class DriverException(Exception): pass people = [('James', 17), ('Kirk', 9), ('Lars', 13), ('Robert', 8)] for person, age in people: if age >= 18: driver = (person, age) break else: raise DriverException('Driver not found.')
Notice that we aren't forced to use the flag
pattern any more. The exception is raised as part of the for
loop logic, which makes good sense because the for
loop is checking on some condition. All we need is to set up a driver
object in case we find one, because the rest of the code is going to use that information somewhere. Notice the code is shorter and more elegant, because the logic is now correctly grouped together where it belongs.
The while loop
In the preceding pages, we saw the for
loop in action. It's incredibly useful when you need to loop over a sequence or a collection. The key point to keep in mind, when you need to be able to discriminate which looping construct to use, is that the for
loop rocks when you have to iterate over a finite amount of elements. It can be a huge amount, but still, something that at some point ends.
There are other cases though, when you just need to loop until some condition is satisfied, or even loop indefinitely until the application is stopped. Cases where we don't really have something to iterate on, and therefore the for
loop would be a poor choice. But fear not, for these cases Python provides us with the while
loop.
The while
loop is similar to the for
loop, in that they both loop and at each iteration they execute a body of instructions. What is different between them is that the while
loop doesn't loop over a sequence (it can, but you have to manually write the logic and it wouldn't make any sense, you would just want to use a for
loop), rather, it loops as long as a certain condition is satisfied. When the condition is no longer satisfied, the loop ends.
As usual, let's see an example which will clarify everything for us. We want to print the binary representation of a positive number. In order to do so, we repeatedly divide the number by two, collecting the remainder, and then produce the inverse of the list of remainders. Let me give you a small example using number 6, which is 110 in binary.
6 / 2 = 3 (remainder: 0) 3 / 2 = 1 (remainder: 1) 1 / 2 = 0 (remainder: 1) List of remainders: 0, 1, 1. Inverse is 1, 1, 0, which is also the binary representation of 6: 110
Let's write some code to calculate the binary representation for number 39: 1001112.
binary.py
n = 39 remainders = [] while n > 0: remainder = n % 2 # remainder of division by 2 remainders.append(remainder) # we keep track of remainders n //= 2 # we divide n by 2 # reassign the list to its reversed copy and print it remainders = remainders[::-1] print(remainders)
In the preceding code, I highlighted two things: n > 0
, which is the condition to keep looping, and remainders[::-1]
which is a nice and easy way to get the reversed version of a list (missing start
and end
parameters, step = -1
, produces the same list, from end
to start
, in reverse order). We can make the code a little shorter (and more Pythonic), by using the divmod
function, which is called with a number and a divisor, and returns a tuple with the result of the integer division and its remainder. For example, divmod(13, 5)
would return (2, 3)
, and indeed 5 * 2 + 3 = 13.
binary.2.py
n = 39
remainders = []
while n > 0:
n, remainder = divmod(n, 2)
remainders.append(remainder)
# reassign the list to its reversed copy and print it
remainders = remainders[::-1]
print(remainders)
In the preceding code, we have reassigned n to the result of the division by 2, and the remainder, in one single line.
Notice that the condition in a while
loop is a condition to continue looping. If it evaluates to True
, then the body is executed and then another evaluation follows, and so on, until the condition evaluates to False
. When that happens, the loop is exited immediately without executing its body.
Note
If the condition never evaluates to False
, the loop becomes a so called infinite loop. Infinite loops are used for example when polling from network devices: you ask the socket if there is any data, you do something with it if there is any, then you sleep for a small amount of time, and then you ask the socket again, over and over again, without ever stopping.
Having the ability to loop over a condition, or to loop indefinitely, is the reason why the for
loop alone is not enough, and therefore Python provides the while
loop.
Just for fun, let's adapt one of the examples (multiple.sequences.py
) using the while logic.
multiple.sequences.while.py
people = ['Jonas', 'Julio', 'Mike', 'Mez'] ages = [25, 30, 31, 39] position = 0 while position < len(people): person = people[position] age = ages[position] print(person, age) position += 1
In the preceding code, I have highlighted the initialization, condition, and update of the variable position
, which makes it possible to simulate the equivalent for
loop code by handling the iteration variable manually. Everything that can be done with a for
loop can also be done with a while
loop, even though you can see there's a bit of boilerplate you have to go through in order to achieve the same result. The opposite is also true, but simulating a never ending while
loop using a for
loop requires some real trickery, so why would you do that? Use the right tool for the job, and 99.9% of the times you'll be fine.
So, to recap, use a for
loop when you need to iterate over one (or a combination of) iterable, and a while
loop when you need to loop according to a condition being satisfied or not. If you keep in mind the difference between the two purposes, you will never choose the wrong looping construct.
Let's now see how to alter the normal flow of a loop.
According to the task at hand, sometimes you will need to alter the regular flow of a loop. You can either skip a single iteration (as many times you want), or you can break out of the loop entirely. A common use case for skipping iterations is for example when you're iterating over a list of items and you need to work on each of them only if some condition is verified. On the other hand, if you're iterating over a collection of items, and you have found one of them that satisfies some need you have, you may decide not to continue the loop entirely and therefore break out of it. There are countless possible scenarios, so it's better to see a couple of examples.
Let's say you want to apply a 20% discount to all products in a basket list for those which have an expiration date of today. The way you achieve this is to use the continue statement, which tells the looping construct (for
or while
) to immediately stop execution of the body and go to the next iteration, if any. This example will take us a little deeper down the rabbit whole, so be ready to jump.
discount.py
from datetime import date, timedelta today = date.today() tomorrow = today + timedelta(days=1) # today + 1 day is tomorrow products = [ {'sku': '1', 'expiration_date': today, 'price': 100.0}, {'sku': '2', 'expiration_date': tomorrow, 'price': 50}, {'sku': '3', 'expiration_date': today, 'price': 20}, ] for product in products: if product['expiration_date'] != today: continue product['price'] *= 0.8 # equivalent to applying 20% discount print( 'Price for sku', product['sku'], 'is now', product['price'])
You see we start by importing the date
and timedelta
objects, then we set up our products. Those with sku 1
and 3
have an expiration date of today
, which means we want to apply 20% discount on them. We loop over each product
and we inspect the expiration date. If it is not (inequality operator, !=
) today
, we don't want to execute the rest of the body suite, so we continue
.
Notice that is not important where in the body suite you place the continue
statement (you can even use it more than once). When you reach it, execution stops and goes back to the next iteration. If we run the discount.py
module, this is the output:
$ python discount.py Price for sku 1 is now 80.0 Price for sku 3 is now 16.0
Which shows you that the last two lines of the body haven't been executed for sku number 2.
Let's now see an example of breaking out of a loop. Say we want to tell if at least any of the elements in a list evaluates to True
when fed to the bool
function. Given that we need to know if there is at least one, when we find it we don't need to keep scanning the list any further. In Python code, this translates to using the break statement. Let's write this down into code:
any.py
items = [0, None, 0.0, True, 0, 7] # True and 7 evaluate to True found = False # this is called "flag" for item in items: print('scanning item', item) if item: found = True # we update the flag break if found: # we inspect the flag print('At least one item evaluates to True') else: print('All items evaluate to False')
The preceding code is such a common pattern in programming, you will see it a lot. When you inspect items this way, basically what you do is to set up a flag
variable, then start the inspection. If you find one element that matches your criteria (in this example, that evaluates to True
), then you update the flag and stop iterating. After iteration, you inspect the flag and take action accordingly. Execution yields:
$ python any.py scanning item 0 scanning item None scanning item 0.0 scanning item True At least one item evaluates to True
See how execution stopped after True
was found?
The break
statement acts exactly like the continue
one, in that it stops executing the body of the loop immediately, but also, prevents any other iteration to run, effectively breaking out of the loop.
The continue
and break
statements can be used together with no limitation in their number, both in the for
and while
looping constructs.
One of the features I've seen only in the Python language is the ability to have else
clauses after while
and for
loops. It's very rarely used, but it's definitely nice to have. In short, you can have an else
suite after a for
or while
loop. If the loop ends normally, because of exhaustion of the iterator (for
loop) or because the condition is finally not met (while
loop), then the else
suite (if present) is executed. In case execution is interrupted by a break
statement, the else
clause is not executed. Let's take an example of a for
loop that iterates over a group of items, looking for one that would match some condition. In case we don't find at least one that satisfies the condition, we want to raise an exception. This means we want to arrest the regular execution of the program and signal that there was an error, or exception, that we cannot deal with. Exceptions will be the subject of Chapter 7, Testing, Profiling, and Dealing with Exceptions, so don't worry if you don't fully understand them now. Just bear in mind that they will alter the regular flow of the code. Let me now show you two examples that do exactly the same thing, but one of them is using the special for
... else
syntax. Say that we want to find among a collection of people one that could drive a car.
for.no.else.py
class DriverException(Exception): pass people = [('James', 17), ('Kirk', 9), ('Lars', 13), ('Robert', 8)] driver = None for person, age in people: if age >= 18: driver = (person, age) break if driver is None: raise DriverException('Driver not found.')
Notice the flag
pattern again. We set driver to be None
, then if we find one we update the driver
flag, and then, at the end of the loop, we inspect it to see if one was found. I kind of have the feeling that those kids would drive a very metallic car, but anyway, notice that if a driver is not found, a DriverException
is raised, signaling the program that execution cannot continue (we're lacking the driver).
The same functionality can be rewritten a bit more elegantly using the following code:
for.else.py
class DriverException(Exception): pass people = [('James', 17), ('Kirk', 9), ('Lars', 13), ('Robert', 8)] for person, age in people: if age >= 18: driver = (person, age) break else: raise DriverException('Driver not found.')
Notice that we aren't forced to use the flag
pattern any more. The exception is raised as part of the for
loop logic, which makes good sense because the for
loop is checking on some condition. All we need is to set up a driver
object in case we find one, because the rest of the code is going to use that information somewhere. Notice the code is shorter and more elegant, because the logic is now correctly grouped together where it belongs.
The break and continue statements
According to the task at hand, sometimes you will need to alter the regular flow of a loop. You can either skip a single iteration (as many times you want), or you can break out of the loop entirely. A common use case for skipping iterations is for example when you're iterating over a list of items and you need to work on each of them only if some condition is verified. On the other hand, if you're iterating over a collection of items, and you have found one of them that satisfies some need you have, you may decide not to continue the loop entirely and therefore break out of it. There are countless possible scenarios, so it's better to see a couple of examples.
Let's say you want to apply a 20% discount to all products in a basket list for those which have an expiration date of today. The way you achieve this is to use the continue statement, which tells the looping construct (for
or while
) to immediately stop execution of the body and go to the next iteration, if any. This example will take us a little deeper down the rabbit whole, so be ready to jump.
discount.py
from datetime import date, timedelta today = date.today() tomorrow = today + timedelta(days=1) # today + 1 day is tomorrow products = [ {'sku': '1', 'expiration_date': today, 'price': 100.0}, {'sku': '2', 'expiration_date': tomorrow, 'price': 50}, {'sku': '3', 'expiration_date': today, 'price': 20}, ] for product in products: if product['expiration_date'] != today: continue product['price'] *= 0.8 # equivalent to applying 20% discount print( 'Price for sku', product['sku'], 'is now', product['price'])
You see we start by importing the date
and timedelta
objects, then we set up our products. Those with sku 1
and 3
have an expiration date of today
, which means we want to apply 20% discount on them. We loop over each product
and we inspect the expiration date. If it is not (inequality operator, !=
) today
, we don't want to execute the rest of the body suite, so we continue
.
Notice that is not important where in the body suite you place the continue
statement (you can even use it more than once). When you reach it, execution stops and goes back to the next iteration. If we run the discount.py
module, this is the output:
$ python discount.py Price for sku 1 is now 80.0 Price for sku 3 is now 16.0
Which shows you that the last two lines of the body haven't been executed for sku number 2.
Let's now see an example of breaking out of a loop. Say we want to tell if at least any of the elements in a list evaluates to True
when fed to the bool
function. Given that we need to know if there is at least one, when we find it we don't need to keep scanning the list any further. In Python code, this translates to using the break statement. Let's write this down into code:
any.py
items = [0, None, 0.0, True, 0, 7] # True and 7 evaluate to True found = False # this is called "flag" for item in items: print('scanning item', item) if item: found = True # we update the flag break if found: # we inspect the flag print('At least one item evaluates to True') else: print('All items evaluate to False')
The preceding code is such a common pattern in programming, you will see it a lot. When you inspect items this way, basically what you do is to set up a flag
variable, then start the inspection. If you find one element that matches your criteria (in this example, that evaluates to True
), then you update the flag and stop iterating. After iteration, you inspect the flag and take action accordingly. Execution yields:
$ python any.py scanning item 0 scanning item None scanning item 0.0 scanning item True At least one item evaluates to True
See how execution stopped after True
was found?
The break
statement acts exactly like the continue
one, in that it stops executing the body of the loop immediately, but also, prevents any other iteration to run, effectively breaking out of the loop.
The continue
and break
statements can be used together with no limitation in their number, both in the for
and while
looping constructs.
One of the features I've seen only in the Python language is the ability to have else
clauses after while
and for
loops. It's very rarely used, but it's definitely nice to have. In short, you can have an else
suite after a for
or while
loop. If the loop ends normally, because of exhaustion of the iterator (for
loop) or because the condition is finally not met (while
loop), then the else
suite (if present) is executed. In case execution is interrupted by a break
statement, the else
clause is not executed. Let's take an example of a for
loop that iterates over a group of items, looking for one that would match some condition. In case we don't find at least one that satisfies the condition, we want to raise an exception. This means we want to arrest the regular execution of the program and signal that there was an error, or exception, that we cannot deal with. Exceptions will be the subject of Chapter 7, Testing, Profiling, and Dealing with Exceptions, so don't worry if you don't fully understand them now. Just bear in mind that they will alter the regular flow of the code. Let me now show you two examples that do exactly the same thing, but one of them is using the special for
... else
syntax. Say that we want to find among a collection of people one that could drive a car.
for.no.else.py
class DriverException(Exception): pass people = [('James', 17), ('Kirk', 9), ('Lars', 13), ('Robert', 8)] driver = None for person, age in people: if age >= 18: driver = (person, age) break if driver is None: raise DriverException('Driver not found.')
Notice the flag
pattern again. We set driver to be None
, then if we find one we update the driver
flag, and then, at the end of the loop, we inspect it to see if one was found. I kind of have the feeling that those kids would drive a very metallic car, but anyway, notice that if a driver is not found, a DriverException
is raised, signaling the program that execution cannot continue (we're lacking the driver).
The same functionality can be rewritten a bit more elegantly using the following code:
for.else.py
class DriverException(Exception): pass people = [('James', 17), ('Kirk', 9), ('Lars', 13), ('Robert', 8)] for person, age in people: if age >= 18: driver = (person, age) break else: raise DriverException('Driver not found.')
Notice that we aren't forced to use the flag
pattern any more. The exception is raised as part of the for
loop logic, which makes good sense because the for
loop is checking on some condition. All we need is to set up a driver
object in case we find one, because the rest of the code is going to use that information somewhere. Notice the code is shorter and more elegant, because the logic is now correctly grouped together where it belongs.
A special else clause
One of the features I've seen only in the Python language is the ability to have else
clauses after while
and for
loops. It's very rarely used, but it's definitely nice to have. In short, you can have an else
suite after a for
or while
loop. If the loop ends normally, because of exhaustion of the iterator (for
loop) or because the condition is finally not met (while
loop), then the else
suite (if present) is executed. In case execution is interrupted by a break
statement, the else
clause is not executed. Let's take an example of a for
loop that iterates over a group of items, looking for one that would match some condition. In case we don't find at least one that satisfies the condition, we want to raise an exception. This means we want to arrest the regular execution of the program and signal that there was an error, or exception, that we cannot deal with. Exceptions will be the subject of Chapter 7, Testing, Profiling, and Dealing with Exceptions, so don't worry if you don't fully understand them now. Just bear in mind that they will alter the regular flow of the code. Let me now show you two examples that do exactly the same thing, but one of them is using the special for
... else
syntax. Say that we want to find among a collection of people one that could drive a car.
for.no.else.py
class DriverException(Exception): pass people = [('James', 17), ('Kirk', 9), ('Lars', 13), ('Robert', 8)] driver = None for person, age in people: if age >= 18: driver = (person, age) break if driver is None: raise DriverException('Driver not found.')
Notice the flag
pattern again. We set driver to be None
, then if we find one we update the driver
flag, and then, at the end of the loop, we inspect it to see if one was found. I kind of have the feeling that those kids would drive a very metallic car, but anyway, notice that if a driver is not found, a DriverException
is raised, signaling the program that execution cannot continue (we're lacking the driver).
The same functionality can be rewritten a bit more elegantly using the following code:
for.else.py
class DriverException(Exception): pass people = [('James', 17), ('Kirk', 9), ('Lars', 13), ('Robert', 8)] for person, age in people: if age >= 18: driver = (person, age) break else: raise DriverException('Driver not found.')
Notice that we aren't forced to use the flag
pattern any more. The exception is raised as part of the for
loop logic, which makes good sense because the for
loop is checking on some condition. All we need is to set up a driver
object in case we find one, because the rest of the code is going to use that information somewhere. Notice the code is shorter and more elegant, because the logic is now correctly grouped together where it belongs.
Now that you have seen all there is to see about conditionals and loops, it's time to spice things up a little, and see those two examples I anticipated at the beginning of this chapter. We'll mix and match here, so you can see how one can use all these concepts together. Let's start by writing some code to generate a list of prime numbers up to some limit. Please bear in mind that I'm going to write a very inefficient and rudimentary algorithm to detect primes. The important thing for you is to concentrate on those bits in the code that belong to this chapter's subject.
"A prime number (or a prime) is a natural number greater than 1 that has no positive divisors other than 1 and itself. A natural number greater than 1 that is not a prime number is called a composite number."
Based on this definition, if we consider the first 10 natural numbers, we can see that 2, 3, 5, and 7 are primes, while 1, 4, 6, 8, 9, 10 are not. In order to have a computer tell you if a number N is prime, you can divide that number by all natural numbers in the range [2, N). If any of those divisions yields zero as a remainder, then the number is not a prime. Enough chatter, let's get down to business. I'll write two versions of this, the second of which will exploit the for
... else
syntax.
primes.py
primes = [] # this will contain the primes in the end upto = 100 # the limit, inclusive for n in range(2, upto + 1): is_prime = True # flag, new at each iteration of outer for for divisor in range(2, n): if n % divisor == 0: is_prime = False break if is_prime: # check on flag primes.append(n) print(primes)
Lots of things to notice in the preceding code. First of all we set up an empty list primes
, which will contain the primes at the end. The limit is 100, and you can see it's inclusive in the way we call range()
in the outer loop. If we wrote range(2, upto)
that would be [2, upto), right? Therefore range(2, upto + 1)
gives us [2, upto + 1) == [2, upto].
So, two for
loops. In the outer one we loop over the candidate primes, that is, all natural numbers from 2 to upto
. Inside each iteration of this outer loop we set up a flag (which is set to True
at each iteration), and then start dividing the current n
by all numbers from 2 to n – 1. If we find a proper divisor for n
, it means n
is composite, and therefore we set the flag to False
and break the loop. Notice that when we break the inner one, the outer one keeps on going normally. The reason why we break after having found a proper divisor for n
is that we don't need any further information to be able to tell that n
is not a prime.
When we check on the is_prime
flag, if it is still True
, it means we couldn't find any number in [2, n) that is a proper divisor for n
, therefore n
is a prime. We append n
to the primes
list, and hop! Another iteration, until n equals 100.
Running this code yields:
$ python primes.py [2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41, 43, 47, 53, 59, 61, 67, 71, 73, 79, 83, 89, 97]
Before we proceed, one question: of all iterations of the outer loop, one of them is different than all the others. Could you tell which one, and why? Think about it for a second, go back to the code and try to figure it out for yourself, and then keep reading on.
Did you figure it out? If not, don't feel bad, it's perfectly normal. I asked you to do it as a small exercise because it's what coders do all the time. The skill to understand what the code does by simply looking at it is something you build over time. It's very important, so try to exercise it whenever you can. I'll tell you the answer now: the iteration that behaves differently from all others is the first one. The reason is because in the first iteration, n
is 2. Therefore the innermost for
loop won't even run, because it's a for
loop which iterates over range(2, 2)
, and what is that if not [2, 2)? Try it out for yourself, write a simple for
loop with that iterable, put a print
in the body suite, and see if anything happens (it won't...).
Now, from an algorithmic point of view this code is inefficient so let's at least make it more beautiful:
primes.else.py
primes = [] upto = 100 for n in range(2, upto + 1): for divisor in range(2, n): if n % divisor == 0: break else: primes.append(n) print(primes)
Much nicer, right? The is_prime
flag is completely gone, and we append n
to the primes
list when we know the inner for
loop hasn't encountered any break
statements. See how the code looks cleaner and reads better?
In this example, I want to show you a technique I like a lot. In many programming languages, other than the if
/elif
/else
constructs, in whatever form or syntax they may come, you can find another statement, usually called switch
/case
, that in Python is missing. It is the equivalent of a cascade of if
/elif
/.../elif
/else clauses, with a syntax similar to this (warning! JavaScript code!):
switch.js
switch (day_number) { case 1: case 2: case 3: case 4: case 5: day = "Weekday"; break; case 6: day = "Saturday"; break; case 0: day = "Sunday"; break; default: day = ""; alert(day_number + ' is not a valid day number.') }
In the preceding code, we switch
on a variable called day_number
. This means we get its value and then we decide what case it fits in (if any). From 1 to 5 there is a cascade, which means no matter the number, [1, 5] all go down to the bit of logic that sets day
as "Weekday"
. Then we have single cases for 0 and 6 and a default
case to prevent errors, which alerts the system that day_number
is not a valid day number, that is, not in [0, 6]. Python is perfectly capable of realizing such logic using if
/elif
/else
statements:
switch.py
if 1 <= day_number <= 5: day = 'Weekday' elif day_number == 6: day = 'Saturday' elif day_number == 0: day = 'Sunday' else: day = '' raise ValueError( str(day_number) + ' is not a valid day number.')
In the preceding code, we reproduce the same logic of the JavaScript snippet, in Python, using if
/elif
/else
statements. I raised ValueError
exception just as an example at the end, if day_number
is not in [0, 6]. This is one possible way of translating the switch
/case
logic, but there is also another one, sometimes called dispatching, which I will show you in the last version of the next example.
Let's start the new example by simply writing some code that assigns a discount to customers based on their coupon value. I'll keep the logic down to a minimum here, remember that all we really care about is conditionals and loops.
coupons.py
customers = [ dict(id=1, total=200, coupon_code='F20'), # F20: fixed, £20 dict(id=2, total=150, coupon_code='P30'), # P30: percent, 30% dict(id=3, total=100, coupon_code='P50'), # P50: percent, 50% dict(id=4, total=110, coupon_code='F15'), # F15: fixed, £15 ] for customer in customers: code = customer['coupon_code'] if code == 'F20': customer['discount'] = 20.0 elif code == 'F15': customer['discount'] = 15.0 elif code == 'P30': customer['discount'] = customer['total'] * 0.3 elif code == 'P50': customer['discount'] = customer['total'] * 0.5 else: customer['discount'] = 0.0 for customer in customers: print(customer['id'], customer['total'], customer['discount'])
We start by setting up some customers. They have an order total, a coupon code, and an id. I made up four different types of coupon, two are fixed and two are percentage based. You can see that in the if
/elif
/else
cascade I apply the discount accordingly, and I set it as a 'discount'
key in the customer
dict.
At the end I just print out part of the data to see if my code is working properly.
$ python coupons.py 1 200 20.0 2 150 45.0 3 100 50.0 4 110 15.0
This code is simple to understand, but all those clauses are kind of cluttering the logic. It's not easy to see what's going on at a first glance, and I don't like it. In cases like this, you can exploit a dictionary to your advantage, like this:
coupons.dict.py
customers = [ dict(id=1, total=200, coupon_code='F20'), # F20: fixed, £20 dict(id=2, total=150, coupon_code='P30'), # P30: percent, 30% dict(id=3, total=100, coupon_code='P50'), # P50: percent, 50% dict(id=4, total=110, coupon_code='F15'), # F15: fixed, £15 ] discounts = { 'F20': (0.0, 20.0), # each value is (percent, fixed) 'P30': (0.3, 0.0), 'P50': (0.5, 0.0), 'F15': (0.0, 15.0), } for customer in customers: code = customer['coupon_code'] percent, fixed = discounts.get(code, (0.0, 0.0)) customer['discount'] = percent * customer['total'] + fixed for customer in customers: print(customer['id'], customer['total'], customer['discount'])
Running the preceding code yields exactly the same result we had from the snippet before it. We spared two lines, but more importantly, we gained a lot in readability, as the body of the for
loop now is just three lines long, and very easy to understand. The concept here is to use a dictionary as dispatcher. In other words, we try to fetch something from the dictionary based on a code (our coupon_code
), and by using dict.get(key, default)
, we make sure we also cater for when the code
is not in the dictionary and we need a default value.
Notice that I had to apply some very simple linear algebra in order to calculate the discount properly. Each discount has a percentage and fixed part in the dictionary, represented by a 2-tuple. By applying percent * total + fixed
, we get the correct discount. When percent
is 0
, the formula just gives the fixed amount, and it gives percent * total
when fixed is 0
. Simple but effective.
This technique is important because it is also used in other contexts, with functions, where it actually becomes much more powerful than what we've seen in the preceding snippet. If it's not completely clear to you how it works, I suggest you to take your time and experiment with it. Change values and add print statements to see what's going on while the program is running.
Example 1 – a prime generator
"A prime number (or a prime) is a natural number greater than 1 that has no positive divisors other than 1 and itself. A natural number greater than 1 that is not a prime number is called a composite number."
Based on this definition, if we consider the first 10 natural numbers, we can see that 2, 3, 5, and 7 are primes, while 1, 4, 6, 8, 9, 10 are not. In order to have a computer tell you if a number N is prime, you can divide that number by all natural numbers in the range [2, N). If any of those divisions yields zero as a remainder, then the number is not a prime. Enough chatter, let's get down to business. I'll write two versions of this, the second of which will exploit the for
... else
syntax.
primes.py
primes = [] # this will contain the primes in the end upto = 100 # the limit, inclusive for n in range(2, upto + 1): is_prime = True # flag, new at each iteration of outer for for divisor in range(2, n): if n % divisor == 0: is_prime = False break if is_prime: # check on flag primes.append(n) print(primes)
Lots of things to notice in the preceding code. First of all we set up an empty list primes
, which will contain the primes at the end. The limit is 100, and you can see it's inclusive in the way we call range()
in the outer loop. If we wrote range(2, upto)
that would be [2, upto), right? Therefore range(2, upto + 1)
gives us [2, upto + 1) == [2, upto].
So, two for
loops. In the outer one we loop over the candidate primes, that is, all natural numbers from 2 to upto
. Inside each iteration of this outer loop we set up a flag (which is set to True
at each iteration), and then start dividing the current n
by all numbers from 2 to n – 1. If we find a proper divisor for n
, it means n
is composite, and therefore we set the flag to False
and break the loop. Notice that when we break the inner one, the outer one keeps on going normally. The reason why we break after having found a proper divisor for n
is that we don't need any further information to be able to tell that n
is not a prime.
When we check on the is_prime
flag, if it is still True
, it means we couldn't find any number in [2, n) that is a proper divisor for n
, therefore n
is a prime. We append n
to the primes
list, and hop! Another iteration, until n equals 100.
Running this code yields:
$ python primes.py [2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41, 43, 47, 53, 59, 61, 67, 71, 73, 79, 83, 89, 97]
Before we proceed, one question: of all iterations of the outer loop, one of them is different than all the others. Could you tell which one, and why? Think about it for a second, go back to the code and try to figure it out for yourself, and then keep reading on.
Did you figure it out? If not, don't feel bad, it's perfectly normal. I asked you to do it as a small exercise because it's what coders do all the time. The skill to understand what the code does by simply looking at it is something you build over time. It's very important, so try to exercise it whenever you can. I'll tell you the answer now: the iteration that behaves differently from all others is the first one. The reason is because in the first iteration, n
is 2. Therefore the innermost for
loop won't even run, because it's a for
loop which iterates over range(2, 2)
, and what is that if not [2, 2)? Try it out for yourself, write a simple for
loop with that iterable, put a print
in the body suite, and see if anything happens (it won't...).
Now, from an algorithmic point of view this code is inefficient so let's at least make it more beautiful:
primes.else.py
primes = [] upto = 100 for n in range(2, upto + 1): for divisor in range(2, n): if n % divisor == 0: break else: primes.append(n) print(primes)
Much nicer, right? The is_prime
flag is completely gone, and we append n
to the primes
list when we know the inner for
loop hasn't encountered any break
statements. See how the code looks cleaner and reads better?
In this example, I want to show you a technique I like a lot. In many programming languages, other than the if
/elif
/else
constructs, in whatever form or syntax they may come, you can find another statement, usually called switch
/case
, that in Python is missing. It is the equivalent of a cascade of if
/elif
/.../elif
/else clauses, with a syntax similar to this (warning! JavaScript code!):
switch.js
switch (day_number) { case 1: case 2: case 3: case 4: case 5: day = "Weekday"; break; case 6: day = "Saturday"; break; case 0: day = "Sunday"; break; default: day = ""; alert(day_number + ' is not a valid day number.') }
In the preceding code, we switch
on a variable called day_number
. This means we get its value and then we decide what case it fits in (if any). From 1 to 5 there is a cascade, which means no matter the number, [1, 5] all go down to the bit of logic that sets day
as "Weekday"
. Then we have single cases for 0 and 6 and a default
case to prevent errors, which alerts the system that day_number
is not a valid day number, that is, not in [0, 6]. Python is perfectly capable of realizing such logic using if
/elif
/else
statements:
switch.py
if 1 <= day_number <= 5: day = 'Weekday' elif day_number == 6: day = 'Saturday' elif day_number == 0: day = 'Sunday' else: day = '' raise ValueError( str(day_number) + ' is not a valid day number.')
In the preceding code, we reproduce the same logic of the JavaScript snippet, in Python, using if
/elif
/else
statements. I raised ValueError
exception just as an example at the end, if day_number
is not in [0, 6]. This is one possible way of translating the switch
/case
logic, but there is also another one, sometimes called dispatching, which I will show you in the last version of the next example.
Let's start the new example by simply writing some code that assigns a discount to customers based on their coupon value. I'll keep the logic down to a minimum here, remember that all we really care about is conditionals and loops.
coupons.py
customers = [ dict(id=1, total=200, coupon_code='F20'), # F20: fixed, £20 dict(id=2, total=150, coupon_code='P30'), # P30: percent, 30% dict(id=3, total=100, coupon_code='P50'), # P50: percent, 50% dict(id=4, total=110, coupon_code='F15'), # F15: fixed, £15 ] for customer in customers: code = customer['coupon_code'] if code == 'F20': customer['discount'] = 20.0 elif code == 'F15': customer['discount'] = 15.0 elif code == 'P30': customer['discount'] = customer['total'] * 0.3 elif code == 'P50': customer['discount'] = customer['total'] * 0.5 else: customer['discount'] = 0.0 for customer in customers: print(customer['id'], customer['total'], customer['discount'])
We start by setting up some customers. They have an order total, a coupon code, and an id. I made up four different types of coupon, two are fixed and two are percentage based. You can see that in the if
/elif
/else
cascade I apply the discount accordingly, and I set it as a 'discount'
key in the customer
dict.
At the end I just print out part of the data to see if my code is working properly.
$ python coupons.py 1 200 20.0 2 150 45.0 3 100 50.0 4 110 15.0
This code is simple to understand, but all those clauses are kind of cluttering the logic. It's not easy to see what's going on at a first glance, and I don't like it. In cases like this, you can exploit a dictionary to your advantage, like this:
coupons.dict.py
customers = [ dict(id=1, total=200, coupon_code='F20'), # F20: fixed, £20 dict(id=2, total=150, coupon_code='P30'), # P30: percent, 30% dict(id=3, total=100, coupon_code='P50'), # P50: percent, 50% dict(id=4, total=110, coupon_code='F15'), # F15: fixed, £15 ] discounts = { 'F20': (0.0, 20.0), # each value is (percent, fixed) 'P30': (0.3, 0.0), 'P50': (0.5, 0.0), 'F15': (0.0, 15.0), } for customer in customers: code = customer['coupon_code'] percent, fixed = discounts.get(code, (0.0, 0.0)) customer['discount'] = percent * customer['total'] + fixed for customer in customers: print(customer['id'], customer['total'], customer['discount'])
Running the preceding code yields exactly the same result we had from the snippet before it. We spared two lines, but more importantly, we gained a lot in readability, as the body of the for
loop now is just three lines long, and very easy to understand. The concept here is to use a dictionary as dispatcher. In other words, we try to fetch something from the dictionary based on a code (our coupon_code
), and by using dict.get(key, default)
, we make sure we also cater for when the code
is not in the dictionary and we need a default value.
Notice that I had to apply some very simple linear algebra in order to calculate the discount properly. Each discount has a percentage and fixed part in the dictionary, represented by a 2-tuple. By applying percent * total + fixed
, we get the correct discount. When percent
is 0
, the formula just gives the fixed amount, and it gives percent * total
when fixed is 0
. Simple but effective.
This technique is important because it is also used in other contexts, with functions, where it actually becomes much more powerful than what we've seen in the preceding snippet. If it's not completely clear to you how it works, I suggest you to take your time and experiment with it. Change values and add print statements to see what's going on while the program is running.
Example 2 – applying discounts
In this example, I want to show you a technique I like a lot. In many programming languages, other than the if
/elif
/else
constructs, in whatever form or syntax they may come, you can find another statement, usually called switch
/case
, that in Python is missing. It is the equivalent of a cascade of if
/elif
/.../elif
/else clauses, with a syntax similar to this (warning! JavaScript code!):
switch.js
switch (day_number) { case 1: case 2: case 3: case 4: case 5: day = "Weekday"; break; case 6: day = "Saturday"; break; case 0: day = "Sunday"; break; default: day = ""; alert(day_number + ' is not a valid day number.') }
In the preceding code, we switch
on a variable called day_number
. This means we get its value and then we decide what case it fits in (if any). From 1 to 5 there is a cascade, which means no matter the number, [1, 5] all go down to the bit of logic that sets day
as "Weekday"
. Then we have single cases for 0 and 6 and a default
case to prevent errors, which alerts the system that day_number
is not a valid day number, that is, not in [0, 6]. Python is perfectly capable of realizing such logic using if
/elif
/else
statements:
switch.py
if 1 <= day_number <= 5: day = 'Weekday' elif day_number == 6: day = 'Saturday' elif day_number == 0: day = 'Sunday' else: day = '' raise ValueError( str(day_number) + ' is not a valid day number.')
In the preceding code, we reproduce the same logic of the JavaScript snippet, in Python, using if
/elif
/else
statements. I raised ValueError
exception just as an example at the end, if day_number
is not in [0, 6]. This is one possible way of translating the switch
/case
logic, but there is also another one, sometimes called dispatching, which I will show you in the last version of the next example.
Let's start the new example by simply writing some code that assigns a discount to customers based on their coupon value. I'll keep the logic down to a minimum here, remember that all we really care about is conditionals and loops.
coupons.py
customers = [ dict(id=1, total=200, coupon_code='F20'), # F20: fixed, £20 dict(id=2, total=150, coupon_code='P30'), # P30: percent, 30% dict(id=3, total=100, coupon_code='P50'), # P50: percent, 50% dict(id=4, total=110, coupon_code='F15'), # F15: fixed, £15 ] for customer in customers: code = customer['coupon_code'] if code == 'F20': customer['discount'] = 20.0 elif code == 'F15': customer['discount'] = 15.0 elif code == 'P30': customer['discount'] = customer['total'] * 0.3 elif code == 'P50': customer['discount'] = customer['total'] * 0.5 else: customer['discount'] = 0.0 for customer in customers: print(customer['id'], customer['total'], customer['discount'])
We start by setting up some customers. They have an order total, a coupon code, and an id. I made up four different types of coupon, two are fixed and two are percentage based. You can see that in the if
/elif
/else
cascade I apply the discount accordingly, and I set it as a 'discount'
key in the customer
dict.
At the end I just print out part of the data to see if my code is working properly.
$ python coupons.py 1 200 20.0 2 150 45.0 3 100 50.0 4 110 15.0
This code is simple to understand, but all those clauses are kind of cluttering the logic. It's not easy to see what's going on at a first glance, and I don't like it. In cases like this, you can exploit a dictionary to your advantage, like this:
coupons.dict.py
customers = [ dict(id=1, total=200, coupon_code='F20'), # F20: fixed, £20 dict(id=2, total=150, coupon_code='P30'), # P30: percent, 30% dict(id=3, total=100, coupon_code='P50'), # P50: percent, 50% dict(id=4, total=110, coupon_code='F15'), # F15: fixed, £15 ] discounts = { 'F20': (0.0, 20.0), # each value is (percent, fixed) 'P30': (0.3, 0.0), 'P50': (0.5, 0.0), 'F15': (0.0, 15.0), } for customer in customers: code = customer['coupon_code'] percent, fixed = discounts.get(code, (0.0, 0.0)) customer['discount'] = percent * customer['total'] + fixed for customer in customers: print(customer['id'], customer['total'], customer['discount'])
Running the preceding code yields exactly the same result we had from the snippet before it. We spared two lines, but more importantly, we gained a lot in readability, as the body of the for
loop now is just three lines long, and very easy to understand. The concept here is to use a dictionary as dispatcher. In other words, we try to fetch something from the dictionary based on a code (our coupon_code
), and by using dict.get(key, default)
, we make sure we also cater for when the code
is not in the dictionary and we need a default value.
Notice that I had to apply some very simple linear algebra in order to calculate the discount properly. Each discount has a percentage and fixed part in the dictionary, represented by a 2-tuple. By applying percent * total + fixed
, we get the correct discount. When percent
is 0
, the formula just gives the fixed amount, and it gives percent * total
when fixed is 0
. Simple but effective.
This technique is important because it is also used in other contexts, with functions, where it actually becomes much more powerful than what we've seen in the preceding snippet. If it's not completely clear to you how it works, I suggest you to take your time and experiment with it. Change values and add print statements to see what's going on while the program is running.
A chapter about iterables, iterators, conditional logic, and looping wouldn't be complete without spending a few words about the itertools
module. If you are into iterating, this is a kind of heaven.
According to the Python official documentation, the itertools
module is:
"A module which implements a number of iterator building blocks inspired by constructs from APL, Haskell, and SML. Each has been recast in a form suitable for Python. The module standardizes a core set of fast, memory efficient tools that are useful by themselves or in combination. Together, they form an "iterator algebra" making it possible to construct specialized tools succinctly and efficiently in pure Python."
By no means do I have the room here to show you all the goodies you can find in this module, so I encourage you to go and check it out for yourself, I promise you'll enjoy it.
In a nutshell, it provides you with three broad categories of iterators. I will give you a very small example of one iterator taken from each one of them, just to make your mouth water a little.
Infinite iterators allow you to work with a for
loop in a different fashion, like if it was a while
loop.
infinite.py
from itertools import count for n in count(5, 3): if n > 20: break print(n, end=', ') # instead of newline, comma and space
Running the code gives this:
$ python infinite.py 5, 8, 11, 14, 17, 20,
The count
factory class makes an iterator that just goes on and on counting. It starts from 5 and keeps adding 3 to it. We need to manually break it if we don't want to get stuck in an infinite loop.
This category is very interesting. It allows you to create an iterator based on multiple iterators, combining their values according to some logic. The key point here is that among those iterators, in case any of them are shorter than the rest, the resulting iterator won't break, it will simply stop as soon as the shortest iterator is exhausted. This is very theoretical, I know, so let me give you an example using compress
. This iterator gives you back the data according to a corresponding item in a selector being True
or False
:
compress('ABC', (1, 0, 1))
would give back 'A'
and 'C'
, because they correspond to the 1's
. Let's see a simple example:
compress.py
from itertools import compress data = range(10) even_selector = [1, 0] * 10 odd_selector = [0, 1] * 10 even_numbers = list(compress(data, even_selector)) odd_numbers = list(compress(data, odd_selector)) print(odd_selector) print(list(data)) print(even_numbers) print(odd_numbers)
Notice that odd_selector
and even_selector
are 20 elements long, while data
is just 10 elements long. compress
will stop as soon as data
has yielded its last element. Running this code produces the following:
$ python compress.py [0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1] [0, 1, 2, 3, 4, 5, 6, 7, 8, 9] [0, 2, 4, 6, 8] [1, 3, 5, 7, 9]
It's a very fast and nice way of selecting elements out of an iterable. The code is very simple, just notice that instead of using a for
loop to iterate over each value that is given back by the compress calls, we used list()
, which does the same, but instead of executing a body of instructions, puts all the values into a list and returns it.
Last but not least, combinatoric generators. These are really fun, if you are into this kind of thing. Let's just see a simple example on permutations.
According to Wolfram Mathworld:
"A permutation, also called an "arrangement number" or "order", is a rearrangement of the elements of an ordered list S into a one-to-one correspondence with S itself."
For example, the permutations of ABC are 6: ABC, ACB, BAC, BCA, CAB, and CBA.
If a set has N elements, then the number of permutations of them is N! (N factorial). For the string ABC the permutations are 3! = 3 * 2 * 1 = 6. Let's do it in Python:
permutations.py
from itertools import permutations
print(list(permutations('ABC')))
This very short snippet of code produces the following result:
$ python permutations.py [('A', 'B', 'C'), ('A', 'C', 'B'), ('B', 'A', 'C'), ('B', 'C', 'A'), ('C', 'A', 'B'), ('C', 'B', 'A')]
Be very careful when you play with permutation. Their number grows at a rate that is proportional to the factorial of the number of the elements you're permuting, and that number can get really big, really fast.
Infinite iterators
Infinite iterators allow you to work with a for
loop in a different fashion, like if it was a while
loop.
infinite.py
from itertools import count for n in count(5, 3): if n > 20: break print(n, end=', ') # instead of newline, comma and space
Running the code gives this:
$ python infinite.py 5, 8, 11, 14, 17, 20,
The count
factory class makes an iterator that just goes on and on counting. It starts from 5 and keeps adding 3 to it. We need to manually break it if we don't want to get stuck in an infinite loop.
This category is very interesting. It allows you to create an iterator based on multiple iterators, combining their values according to some logic. The key point here is that among those iterators, in case any of them are shorter than the rest, the resulting iterator won't break, it will simply stop as soon as the shortest iterator is exhausted. This is very theoretical, I know, so let me give you an example using compress
. This iterator gives you back the data according to a corresponding item in a selector being True
or False
:
compress('ABC', (1, 0, 1))
would give back 'A'
and 'C'
, because they correspond to the 1's
. Let's see a simple example:
compress.py
from itertools import compress data = range(10) even_selector = [1, 0] * 10 odd_selector = [0, 1] * 10 even_numbers = list(compress(data, even_selector)) odd_numbers = list(compress(data, odd_selector)) print(odd_selector) print(list(data)) print(even_numbers) print(odd_numbers)
Notice that odd_selector
and even_selector
are 20 elements long, while data
is just 10 elements long. compress
will stop as soon as data
has yielded its last element. Running this code produces the following:
$ python compress.py [0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1] [0, 1, 2, 3, 4, 5, 6, 7, 8, 9] [0, 2, 4, 6, 8] [1, 3, 5, 7, 9]
It's a very fast and nice way of selecting elements out of an iterable. The code is very simple, just notice that instead of using a for
loop to iterate over each value that is given back by the compress calls, we used list()
, which does the same, but instead of executing a body of instructions, puts all the values into a list and returns it.
Last but not least, combinatoric generators. These are really fun, if you are into this kind of thing. Let's just see a simple example on permutations.
According to Wolfram Mathworld:
"A permutation, also called an "arrangement number" or "order", is a rearrangement of the elements of an ordered list S into a one-to-one correspondence with S itself."
For example, the permutations of ABC are 6: ABC, ACB, BAC, BCA, CAB, and CBA.
If a set has N elements, then the number of permutations of them is N! (N factorial). For the string ABC the permutations are 3! = 3 * 2 * 1 = 6. Let's do it in Python:
permutations.py
from itertools import permutations
print(list(permutations('ABC')))
This very short snippet of code produces the following result:
$ python permutations.py [('A', 'B', 'C'), ('A', 'C', 'B'), ('B', 'A', 'C'), ('B', 'C', 'A'), ('C', 'A', 'B'), ('C', 'B', 'A')]
Be very careful when you play with permutation. Their number grows at a rate that is proportional to the factorial of the number of the elements you're permuting, and that number can get really big, really fast.
Iterators terminating on the shortest input sequence
This category is very interesting. It allows you to create an iterator based on multiple iterators, combining their values according to some logic. The key point here is that among those iterators, in case any of them are shorter than the rest, the resulting iterator won't break, it will simply stop as soon as the shortest iterator is exhausted. This is very theoretical, I know, so let me give you an example using compress
. This iterator gives you back the data according to a corresponding item in a selector being True
or False
:
compress('ABC', (1, 0, 1))
would give back 'A'
and 'C'
, because they correspond to the 1's
. Let's see a simple example:
compress.py
from itertools import compress data = range(10) even_selector = [1, 0] * 10 odd_selector = [0, 1] * 10 even_numbers = list(compress(data, even_selector)) odd_numbers = list(compress(data, odd_selector)) print(odd_selector) print(list(data)) print(even_numbers) print(odd_numbers)
Notice that odd_selector
and even_selector
are 20 elements long, while data
is just 10 elements long. compress
will stop as soon as data
has yielded its last element. Running this code produces the following:
$ python compress.py [0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1] [0, 1, 2, 3, 4, 5, 6, 7, 8, 9] [0, 2, 4, 6, 8] [1, 3, 5, 7, 9]
It's a very fast and nice way of selecting elements out of an iterable. The code is very simple, just notice that instead of using a for
loop to iterate over each value that is given back by the compress calls, we used list()
, which does the same, but instead of executing a body of instructions, puts all the values into a list and returns it.
Last but not least, combinatoric generators. These are really fun, if you are into this kind of thing. Let's just see a simple example on permutations.
According to Wolfram Mathworld:
"A permutation, also called an "arrangement number" or "order", is a rearrangement of the elements of an ordered list S into a one-to-one correspondence with S itself."
For example, the permutations of ABC are 6: ABC, ACB, BAC, BCA, CAB, and CBA.
If a set has N elements, then the number of permutations of them is N! (N factorial). For the string ABC the permutations are 3! = 3 * 2 * 1 = 6. Let's do it in Python:
permutations.py
from itertools import permutations
print(list(permutations('ABC')))
This very short snippet of code produces the following result:
$ python permutations.py [('A', 'B', 'C'), ('A', 'C', 'B'), ('B', 'A', 'C'), ('B', 'C', 'A'), ('C', 'A', 'B'), ('C', 'B', 'A')]
Be very careful when you play with permutation. Their number grows at a rate that is proportional to the factorial of the number of the elements you're permuting, and that number can get really big, really fast.
Combinatoric generators
Last but not least, combinatoric generators. These are really fun, if you are into this kind of thing. Let's just see a simple example on permutations.
According to Wolfram Mathworld:
"A permutation, also called an "arrangement number" or "order", is a rearrangement of the elements of an ordered list S into a one-to-one correspondence with S itself."
For example, the permutations of ABC are 6: ABC, ACB, BAC, BCA, CAB, and CBA.
If a set has N elements, then the number of permutations of them is N! (N factorial). For the string ABC the permutations are 3! = 3 * 2 * 1 = 6. Let's do it in Python:
permutations.py
from itertools import permutations
print(list(permutations('ABC')))
This very short snippet of code produces the following result:
$ python permutations.py [('A', 'B', 'C'), ('A', 'C', 'B'), ('B', 'A', 'C'), ('B', 'C', 'A'), ('C', 'A', 'B'), ('C', 'B', 'A')]
Be very careful when you play with permutation. Their number grows at a rate that is proportional to the factorial of the number of the elements you're permuting, and that number can get really big, really fast.
In this chapter, we've taken another step forward to expand our coding vocabulary. We've seen how to drive the execution of the code by evaluating conditions, and we've seen how to loop and iterate over sequences and collections of objects. This gives us the power to control what happens when our code is run, which means we are getting an idea on how to shape it so that it does what we want and it reacts to data that changes dynamically.
We've also seen how to combine everything together in a couple of simple examples, and in the end we have taken a brief look at the itertools
module, which is full of interesting iterators which can enrich our abilities with Python even more.
Now it's time to switch gears, to take another step forward and talk about functions. The next chapter is all about them because they are extremely important. Make sure you're comfortable with what has been done up to now: I want to provide you with interesting examples, so I'll have to go a little faster. Ready? Turn the page.
"To create architecture is to put in order. Put what in order? Function and objects." | ||
--Le Corbusier |
In this chapter, we're going to explore functions. We already said that everything is an object in Python, and functions are no exception to this. But, what exactly is a function? A function is a sequence of instructions that perform a task, bundled as a unit. This unit can then be imported and used wherever it's needed. There are many advantages to using functions in your code, as we'll see shortly.
I believe the saying, a picture is worth one thousand words, is particularly true when explaining functions to someone who is new to this concept, so please take a look at the following image:

As you can see, a function is a block of instructions, packaged as a whole, like a box. Functions can accept input arguments and produce output values. Both of these are optional, as we'll see in the examples in this chapter.
A function in Python is defined by using the def
keyword, after which the name of the function follows, terminated by a pair of braces (which may or may not contain input parameters) and, finally, a colon (:
) signals the end of the function definition line. Immediately afterwards, indented by four spaces, we find the body of the function, which is the set of instructions that the function will execute when called.
A function may or may not return output. If a function wants to return output, it does so by using the return
keyword, followed by the desired output. If you have an eagle eye, you may have noticed the little * after Optional in the output section of the preceding picture. This is because a function always returns something in Python, even if you don't explicitly use the return
clause. If the function has no return
statement in its body, it's return value is None
. The reasons behind this design choice are out of the scope of an introductory chapter, so all you need to know is that this behavior will make your life easier, as always, thank you Python.
Functions are among the most important concepts and constructs of any language, so let me give you a few reasons why we need them:
- They reduce code duplication in a program. By having a specific task taken care of by a nice block of packaged code that we can import and call whenever we want, we don't need to duplicate its implementation.
- They help in splitting a complex task or procedure into smaller blocks, each of which becomes a function.
- They hide the implementation details from their users.
- They improve traceability.
- They improve readability.
Let's look at a few examples to get a better understanding of each point.
Imagine that you are writing a piece of scientific software, and you need to calculate primes up to a limit, as we did in the previous chapter. You write several algorithms and prime numbers, being the basis of many different types of calculations, keep creeping into your code. Well, you have a nice algorithm to calculate them, so you copy and paste it to wherever you need. One day, though, your friend Mister Smarty gives you a better algorithm to calculate prime numbers, and this will save you a lot of time. At this point, you need to go over your whole codebase and replace the old code with the new code.
This is actually a very bad way to go about it. It's error-prone, you never know what lines you are chopping out or leaving there by mistake when you cut and paste code in other code, and you may also risk missing one of the places where prime calculation was done, leaving your software with different versions. Can you imagine if you discovered that the old way was buggy? You would have an undetected bug in your code, and bugs like this are quite hard to spot, especially in big codebases.
So, what should you do? Simple! You write a function, get_prime_numbers(upto)
, and use it anywhere you need a list of primes. When Mister Smarty comes to you and gives you the new code, all you have to do is replace the body of that function with the new implementation, and you're done! The rest of the software will automatically adapt, since it's just calling the function.
Your code will be shorter, it will not suffer from inconsistencies between old and new ways of performing a task, or undetected bugs due to copy and paste failures or oversights. Use functions, and you'll only gain from it, I promise.
Functions are very useful also to split a long or complex task into smaller pieces. The end result is that the code benefits from it in several ways, for example, readability, testability, and reuse. To give you a simple example, imagine that you're preparing a report. Your code needs to fetch data from a data source, parse it, filter it, polish it, and then a whole series of algorithms needs to be run against it, in order to produce the results which will feed the Report
class. It's not uncommon to read procedures like this that are just one big function do_report(data_source)
. There are tens or hundreds of lines of code which end with return report
.
Situations like this are common in code produced by scientists. They have brilliant minds and they care about the correctness of the end result but, unfortunately, sometimes they have no training in programming theory. It is not their fault, one cannot know everything. Now, picture in your head something like a few hundred lines of code. It's very hard to follow through, to find the places where things are changing context (like finishing one task and starting the next one). Do you have the picture in your mind? Good. Don't do it! Instead, look at this code:
data.science.example.py
def do_report(data_source): # fetch and prepare data data = fetch_data(data_source) parsed_data = parse_data(data) filtered_data = filter_data(parsed_data) polished_data = polish_data(filtered_data) # run algorithms on data final_data = analyse(polished_data) # create and return report report = Report(final_data) return report
The previous example is fictitious, of course, but can you see how easy it would be to go through the code? If the end result looks wrong, it would be very easy to debug each of the single data outputs in the do_report
function. Moreover, it's even easier to exclude part of the process temporarily from the whole procedure (you just need to comment out the parts you need to suspend). Code like this is easier to deal with.
Let's stay with the preceding example to talk about this point as well. You can see that, by going through the code of the do_report
function, you can get a pretty good understanding without reading one single line of implementation. This is because functions hide the implementation details. This feature means that, if you don't need to delve into details, you are not forced to, in the way you would if do_report
was just one big fat function. In order to understand what was going on, you would have to read the implementation details. You don't need to with functions. This reduces the time you spend reading the code and since, in a professional environment, reading code takes much more time than actually writing it, it's very important to reduce it as much as we can.
Coders sometimes don't see the point in writing a function with a body of one or two lines of code, so let's look at an example that shows you why you should do it.
Imagine that you need to multiply two matrices:

Would you prefer to have to read this code:
matrix.multiplication.nofunc.py
a = [[1, 2], [3, 4]] b = [[5, 1], [2, 1]] c = [[sum(i * j for i, j in zip(r, c)) for c in zip(*b)] for r in a]
matrix.multiplication.func.py
# this function could also be defined in another module
def matrix_mul(a, b):
return [[sum(i * j for i, j in zip(r, c)) for c in zip(*b)]
for r in a]
a = [[1, 2], [3, 4]]
b = [[5, 1], [2, 1]]
c = matrix_mul(a, b)
It's much easier to understand that c
is the result of the multiplication between a
and b
in the second example. It's much easier to read through the code and, if you don't need to modify that part, you don't even need to go into the implementation details.
Therefore, readability is improved here while, in the first snippet, you would have to spend time trying to understand what that complicated list comprehension was doing.
Imagine that you have written an e-commerce website. You have displayed the product prices all over the pages. Imagine that the prices in your database are stored with no VAT, but you want to display them on the website with VAT at 20%. Here's a few ways of calculating the VAT-inclusive price from the VAT-exclusive price.
vat.py
price = 100 # GBP, no VAT final_price1 = price * 1.2 final_price2 = price + price / 5.0 final_price3 = price * (100 + 20) / 100.0 final_price4 = price + price * 0.2
All these four different ways of calculating a VAT-inclusive price are perfectly acceptable, and I promise you I have found them all in my colleagues' code, over the years. Now, imagine that you have started selling your products in different countries and some of them have different VAT rates so you need to refactor your code (throughout the website) in order to make that VAT calculation dynamic.
How do you trace all the places in which you are performing a VAT calculation? Coding today is a collaborative task and you cannot be sure the VAT has been calculated using only one of those forms. It's going to be hell, believe me.
So, let's write a function that takes the input values, vat
and price
(VAT-exclusive), and returns a VAT-inclusive price.
vat.function.py
def calculate_price_with_vat(price, vat): return price * (100 + vat) / 100
Now you can import that function and apply it in any place of your website where you need to calculate a VAT-inclusive price and when you need to trace those calls, you can search for calculate_price_with_vat
.
Do you remember when we talked about scopes and namespaces in the first chapter? We're going to expand on that concept now. Finally, we can talk about functions and this will make everything easier to understand. Let's start with a very simple example.
scoping.level.1.py
def my_function(): test = 1 # this is defined in the local scope of the function print('my_function:', test) test = 0 # this is defined in the global scope my_function() print('global:', test)
I have defined the name test
in two different places in the previous example. It is actually in two different scopes. One is the global scope (test = 0
), and the other is the local scope of the function my_function
(test = 1
). If you execute the code, you'll see this:
$ python scoping.level.1.py my_function: 1 global: 0
It's clear that test = 1
shadows the assignment test = 0
in my_function
. In the global context, test
is still 0
, as you can see from the output of the program but we define the name test
again in the function body, and we set it to point to an integer of value 1
. Both the two test
names therefore exist, one in the global scope, pointing to an int
object with value 0, the other in the my_function
scope, pointing to an int
object with value 1. Let's comment out the line with test = 1
. Python goes and searches for the name test
in the next enclosing namespace (recall the LEGB rule: Local, Enclosing, Global, Built-in described in Chapter 1, Introduction and First Steps – Take a Deep Breath) and, in this case, we will see the value 0
printed twice. Try it in your code.
Now, let's raise the stakes here and level up:
scoping.level.2.py
def outer(): test = 1 # outer scope def inner(): test = 2 # inner scope print('inner:', test) inner() print('outer:', test) test = 0 # global scope outer() print('global:', test)
In the preceding code, we have two levels of shadowing. One level is in the function outer
, and the other one is in the function inner
. It is far from rocket science, but it can be tricky. If we run the code, we get:
$ python scoping.level.2.py inner: 2 outer: 1 global: 0
Try commenting out the line test = 1
. What do you think the result will be? Well, when reaching the line print('outer:', test)
, Python will have to look for test
in the next enclosing scope, therefore it will find and print 0
, instead of 1
. Make sure you comment out test = 2
as well, to see if you understand what happens, and if the LEGB rule is clear, before proceeding.
Another thing to note is that Python gives you the ability to define a function in another function. The inner function's name is defined within the namespace of the outer function, exactly as would happen with any other name.
Going back to the preceding example, we can alter what happens to the shadowing of the test name by using one of these two special statements: global
and nonlocal
. As you can see from the previous example, when we define test = 2
in the function inner
, we overwrite test
neither in the function outer
, nor in the global scope. We can get read access to those names if we use them in a nested scope that doesn't define them, but we cannot modify them because, when we write an assignment instruction, we're actually defining a new name in the current scope.
How do we change this behavior? Well, we can use the nonlocal
statement. According to the official documentation:
"The
nonlocal
statement causes the listed identifiers to refer to previously bound variables in the nearest enclosing scope excluding globals."
Let's introduce it in the function inner
, and see what happens:
scoping.level.2.nonlocal.py
def outer(): test = 1 # outer scope def inner(): nonlocal test test = 2 # nearest enclosing scope print('inner:', test) inner() print('outer:', test) test = 0 # global scope outer() print('global:', test)
Notice how in the body of the function inner
I have declared the test
name to be nonlocal
. Running this code produces the following result:
$ python scoping.level.2.nonlocal.py inner: 2 outer: 2 global: 0
Wow, look at that result! It means that, by declaring test
to be nonlocal
in the function inner
, we actually get to bind the name test
to that declared in the function outer
. If we removed the nonlocal
test
line from the function inner
and tried the same trick in the function outer
, we would get a SyntaxError
, because the nonlocal
statement works on enclosing scopes excluding the global one.
Is there a way to get to that test = 0
in the global namespace then? Of course, we just need to use the global
statement. Let's try it.
scoping.level.2.global.py
def outer(): test = 1 # outer scope def inner(): global test test = 2 # global scope print('inner:', test) inner() print('outer:', test) test = 0 # global scope outer() print('global:', test)
Note that we have now declared the name test
to be global
, which will basically bind it to the one we defined in the global namespace (test = 0
). Run the code and you should get the following:
$ python scoping.level.2.global.py inner: 2 outer: 1 global: 2
This shows that the name affected by the assignment test = 2
is now the global
one. This trick would also work in the outer
function because, in this case, we're referring to the global scope. Try it for yourself and see what changes, get comfortable with scopes and name resolution, it's very important.
The global and nonlocal statements
Going back to the preceding example, we can alter what happens to the shadowing of the test name by using one of these two special statements: global
and nonlocal
. As you can see from the previous example, when we define test = 2
in the function inner
, we overwrite test
neither in the function outer
, nor in the global scope. We can get read access to those names if we use them in a nested scope that doesn't define them, but we cannot modify them because, when we write an assignment instruction, we're actually defining a new name in the current scope.
How do we change this behavior? Well, we can use the nonlocal
statement. According to the official documentation:
"The
nonlocal
statement causes the listed identifiers to refer to previously bound variables in the nearest enclosing scope excluding globals."
Let's introduce it in the function inner
, and see what happens:
scoping.level.2.nonlocal.py
def outer(): test = 1 # outer scope def inner(): nonlocal test test = 2 # nearest enclosing scope print('inner:', test) inner() print('outer:', test) test = 0 # global scope outer() print('global:', test)
Notice how in the body of the function inner
I have declared the test
name to be nonlocal
. Running this code produces the following result:
$ python scoping.level.2.nonlocal.py inner: 2 outer: 2 global: 0
Wow, look at that result! It means that, by declaring test
to be nonlocal
in the function inner
, we actually get to bind the name test
to that declared in the function outer
. If we removed the nonlocal
test
line from the function inner
and tried the same trick in the function outer
, we would get a SyntaxError
, because the nonlocal
statement works on enclosing scopes excluding the global one.
Is there a way to get to that test = 0
in the global namespace then? Of course, we just need to use the global
statement. Let's try it.
scoping.level.2.global.py
def outer(): test = 1 # outer scope def inner(): global test test = 2 # global scope print('inner:', test) inner() print('outer:', test) test = 0 # global scope outer() print('global:', test)
Note that we have now declared the name test
to be global
, which will basically bind it to the one we defined in the global namespace (test = 0
). Run the code and you should get the following:
$ python scoping.level.2.global.py inner: 2 outer: 1 global: 2
This shows that the name affected by the assignment test = 2
is now the global
one. This trick would also work in the outer
function because, in this case, we're referring to the global scope. Try it for yourself and see what changes, get comfortable with scopes and name resolution, it's very important.
At the beginning of this chapter, we saw that a function can take input parameters. Before we delve into all possible type of parameters, let's make sure you have a clear understanding of what passing a parameter to a function means. There are three key points to keep in mind:
- Argument passing is nothing more than assigning an object to a local variable name
- Assigning an object to an argument name inside a function doesn't affect the caller
- Changing a mutable object argument in a function affects the caller
Let's look at an example for each of these points.
Take a look at the following code. We declare a name x
in the global scope, then we declare a function func(y)
and we call it, passing x
. I highlighted the call in the code.
key.points.argument.passing.py
x = 3
def func(y):
print(y)
func(x) # prints: 3
When func
is called with x
, what happens is that within its local scope, a name y
is created, and it's pointed to the same object x
is pointing to. This is better clarified by the following picture:

The right part of the preceding picture depicts the state of the program when execution has reached the end, after func
has returned (None
). Take a look at the Frames column, and note that we have two names, x and func, in the global namespace (Global frame), pointing to an int (with a value of three) and to a function object, respectively. Right below it, in the rectangle titled func, we can see the function's local namespace, in which only one name has been defined: y. Because we have called func with x (line 5 in the left part of the picture), y is pointing to the same object that x is pointing to. This is what happens under the hood when an argument is passed to a function. If we had used the name x instead of y in the function definition, things would have been exactly the same (only maybe a bit confusing at first), there would be a local x in the function, and a global x outside, as we saw in the Scopes and name resolution section.
So, in a nutshell, what really happens is that the function creates in its local scope the names defined as arguments and, when we call it, we basically tell Python which objects those names must be pointed towards.
This is something that can be tricky to understand at first, so let's look at an example.
key.points.assignment.py
x = 3
def func(x):
x = 7 # defining a local x, not changing the global one
func(x)
print(x) # prints: 3
In the preceding code, when the line x = 7
is executed, what happens is that within the local scope of the function func
, the name x
is pointed to an integer with value 7, leaving the global x
unaltered.
This is the final point, and it's very important because Python apparently behaves differently with mutables (just apparently though). Let's look at an example:
key.points.mutable.py
x = [1, 2, 3]
def func(x):
x[1] = 42 # this affects the caller!
func(x)
print(x) # prints: [1, 42, 3]
Wow, we actually changed the original object! If you think about it, there is nothing weird in this behavior. The name x
in the function is set to point to the caller object by the function call and within the body of the function, we're not changing x
, in that we're not changing its reference, or, in other words, we are not changing the object x
is pointing to. What we're doing is accessing that object's element at position 1, and changing its value.
Remember point #2: "Assigning an object to an argument name within a function doesn't affect the caller". If that is clear to you, the following code should not be surprising.
key.points.mutable.assignment.py
x = [1, 2, 3] def func(x): x[1] = 42 # this changes the caller! x = 'something else' # this points x to a new string object func(x) print(x) # still prints: [1, 42, 3]
Take a look at the two lines I have highlighted. At first, we just access the caller object again, at position 1, and change its value to number 42. Then, we reassign x
to point to the string 'something else'
. This leaves the caller unaltered, according to point #2, and, in fact, the output is the same as that of the previous snippet.
Take your time to play around with this concept and experiment with prints and calls to the id
function until everything is clear in your mind. This is one of the key aspects of Python and it must be very clear, otherwise you risk introducing subtle bugs into your code.
Now that we have a good understanding of input parameters and how they behave, let's see how we can specify them.
There are five different ways of specifying input parameters. Let's look at them one by one.
Positional arguments are read from left to right and they are the most common type of arguments.
arguments.positional.py
def func(a, b, c): print(a, b, c) func(1, 2, 3) # prints: 1 2 3
There is not much else to say. They can be as numerous as you want and they are assigned by position. In the function call, 1
comes first, 2
comes second and 3
comes third, therefore they are assigned to a
, b
and c
respectively.
Keyword arguments are assigned by keyword using the name=value
syntax.
arguments.keyword.py
def func(a, b, c):
print(a, b, c)
func(a=1, c=2, b=3) # prints: 1 3 2
Keyword arguments act when calling the function instead of respecting the left-to-right positional assignment, k. Keyword arguments are matched by name, even when they don't respect the definition's original position (we'll see that there is a limitation to this behavior later, when we mix and match different types of arguments).
The counterpart of keyword arguments, on the definition side, is default values. The syntax is the same, name=value
, and allows us to not have to provide an argument if we are happy with the given default.
arguments.default.py
def func(a, b=4, c=88):
print(a, b, c)
func(1) # prints: 1 4 88
func(b=5, a=7, c=9) # prints: 7 5 9
func(42, c=9) # prints: 42 4 9
The are two things to notice, which are very important. First of all, you cannot specify a default argument on the left of a positional one. Second, note how in the examples, when an argument is passed without using the argument_name=value
syntax, it must be the first one in the list,, and it is always assigned to a
. Try and scramble those arguments and see what happens. Python error messages are very good at telling you what's wrong. So, for example, if you tried something like this:
func(b=1, c=2, 42) # positional argument after keyword one
You would get the following error:
SyntaxError: non-keyword arg after keyword arg
This informs you that you've called the function incorrectly.
Sometimes you may want to pass a variable number of positional arguments to a function and Python provides you with the ability to do it. Let's look at a very common use case, the minimum
function. This is a function that calculates the minimum of its input values.
arguments.variable.positional.py
def minimum(*n):
# print(n) # n is a tuple
if n: # explained after the code
mn = n[0]
for value in n[1:]:
if value < mn:
mn = value
print(mn)
minimum(1, 3, -7, 9) # n = (1, 3, -7, 9) - prints: -7
minimum() # n = () - prints: nothing
As you can see, when we specify a parameter prepending a *
to its name, we are telling Python that that parameter will be collecting a variable number of positional arguments, according to how the function is called. Within the function, n
is a tuple. Uncomment the print(n)
to see for yourself and play around with it for a bit.
Note
Have you noticed how we checked if n
wasn't empty with a simple if n:
? This is due to the fact that collection objects evaluate to True
when non-empty, and otherwise False
in Python. This is true for tuples, sets, lists, dictionaries, and so on.
One other thing to note is that we may want to throw an error when we call the function with no arguments, instead of silently doing nothing. In this context, we're not concerned about making this function robust, but in understanding variable positional arguments.
Let's make another example to show you two things that, in my experience, are confusing to those who are new to this.
arguments.variable.positional.unpacking.py
def func(*args): print(args) values = (1, 3, -7, 9) func(values) # equivalent to: func((1, 3, -7, 9)) func(*values) # equivalent to: func(1, 3, -7, 9)
Take a good look at the last two lines of the preceding example. In the first one, we call func
with one argument, a four elements tuple. In the second example, by using the *
syntax, we're doing something called unpacking, which means that the four elements tuple is unpacked, and the function is called with four arguments: 1, 3, -7, 9
.
This behavior is part of the magic Python does to allow you to do amazing things when calling functions dynamically.
Variable keyword arguments are very similar to variable positional arguments. The only difference is the syntax (**
instead of *
) and that they are collected in a dictionary. Collection and unpacking work in the same way, so let's look at an example:
arguments.variable.keyword.py
def func(**kwargs): print(kwargs) # All calls equivalent. They print: {'a': 1, 'b': 42} func(a=1, b=42) func(**{'a': 1, 'b': 42}) func(**dict(a=1, b=42))
All the calls are equivalent in the preceding example. You can see that adding a **
in front of the parameter name in the function definition tells Python to use that name to collect a variable number of keyword parameters. On the other hand, when we call the function, we can either pass name=value
arguments explicitly, or unpack a dictionary using the same **
syntax.
The reason why being able to pass a variable number of keyword parameters is so important may not be evident at the moment, so, how about a more realistic example? Let's define a function that connects to a database. We want to connect to a default database by simply calling this function with no parameters. We also want to connect to any other database by passing the function the appropriate arguments. Before you read on, spend a couple of minutes figuring out a solution by yourself.
arguments.variable.db.py
def connect(**options): conn_params = { 'host': options.get('host', '127.0.0.1'), 'port': options.get('port', 5432), 'user': options.get('user', ''), 'pwd': options.get('pwd', ''), } print(conn_params) # we then connect to the db (commented out) # db.connect(**conn_params) connect() connect(host='127.0.0.42', port=5433) connect(port=5431, user='fab', pwd='gandalf')
Note in the function we can prepare a dictionary of connection parameters (conn_params
) in the function using default values as fallback, allowing them to be overwritten if they are provided in the function call. There are better ways to do this with fewer lines of code but we're not concerned with that now. Running the preceding code yields the following result:
$ python arguments.variable.db.py {'host': '127.0.0.1', 'pwd': '', 'user': '', 'port': 5432} {'host': '127.0.0.42', 'pwd': '', 'user': '', 'port': 5433} {'host': '127.0.0.1', 'pwd': 'gandalf', 'user': 'fab', 'port': 5431}
Note the correspondence between the function calls and the output. Note how default values are either there or overridden, according to what was passed to the function.
Python 3 allows for a new type of parameter: the keyword-only parameter. We are going to study them only briefly as their use cases are not that frequent. There are two ways of specifying them, either after the variable positional arguments, or after a bare *. Let's see an example of both.
arguments.keyword.only.py
def kwo(*a, c): print(a, c) kwo(1, 2, 3, c=7) # prints: (1, 2, 3) 7 kwo(c=4) # prints: () 4 # kwo(1, 2) # breaks, invalid syntax, with the following error # TypeError: kwo() missing 1 required keyword-only argument: 'c' def kwo2(a, b=42, *, c): print(a, b, c) kwo2(3, b=7, c=99) # prints: 3 7 99 kwo2(3, c=13) # prints: 3 42 13 # kwo2(3, 23) # breaks, invalid syntax, with the following error # TypeError: kwo2() missing 1 required keyword-only argument: 'c'
As anticipated, the function, kwo
, takes a variable number of positional arguments (a
) and a keyword-only function, c
. The results of the calls are straightforward and you can uncomment the third call to see what error Python returns.
The same applies to the function, kwo2
, which differs from kwo
in that it takes a positional argument a
, a keyword argument b
, and then a keyword-only argument, c
. You can uncomment the third call to see the error.
Now that you know how to specify different types of input parameters, let's see how you can combine them in function definitions.
You can combine input parameters, as long as you follow these ordering rules:
- When defining a function, normal positional arguments come first (
name
), then any default arguments (name=value
), then the variable positional arguments (*name
, or simply*
), then any keyword-only arguments (eithername
orname=value
form is good), then any variable keyword arguments (**name
). - On the other hand, when calling a function, arguments must be given in the following order: positional arguments first (
value
), then any combination of keyword arguments (name=value
), variable positional arguments (*name
), then variable keyword arguments (**name
).
Since this can be a bit tricky when left hanging in the theoretical world, let's look at a couple of quick examples.
arguments.all.py
def func(a, b, c=7, *args, **kwargs): print('a, b, c:', a, b, c) print('args:', args) print('kwargs:', kwargs) func(1, 2, 3, *(5, 7, 9), **{'A': 'a', 'B': 'b'}) func(1, 2, 3, 5, 7, 9, A='a', B='b') # same as previous one
Note the order of the parameters in the function definition, and that the two calls are equivalent. In the first one, we're using the unpacking operators for iterables and dictionaries, while in the second one we're using a more explicit syntax. The execution of this yields (I printed only the result of one call):
$ python arguments.all.py a, b, c: 1 2 3 args: (5, 7, 9) kwargs: {'A': 'a', 'B': 'b'}
Let's now look at an example with keyword-only arguments.
arguments.all.kwonly.py
def func_with_kwonly(a, b=42, *args, c, d=256, **kwargs):
print('a, b:', a, b)
print('c, d:', c, d)
print('args:', args)
print('kwargs:', kwargs)
# both calls equivalent
func_with_kwonly(3, 42, c=0, d=1, *(7, 9, 11), e='E', f='F')
func_with_kwonly(3, 42, *(7, 9, 11), c=0, d=1, e='E', f='F')
Note that I have highlighted the keyword-only arguments in the function declaration. They come after the variable positional argument *args
, and it would be the same if they came right after a single *
(in which case there wouldn't be a variable positional argument). The execution of this yields (I printed only the result of one call):
$ python arguments.all.kwonly.py a, b: 3 42 c, d: 0 1 args: (7, 9, 11) kwargs: {'f': 'F', 'e': 'E'}
One other thing to note are the names I gave to the variable positional and keyword arguments. You're free to choose differently, but be aware that args
and kwargs
are the conventional names given to these parameters, at least generically. Now that you know how to define a function in all possible flavors, let me show you something tricky: mutable defaults.
One thing to be very aware of with Python is that default values are created at def
time, therefore, subsequent calls to the same function will possibly behave differently according to the mutability of their default values. Let's look at an example:
arguments.defaults.mutable.py
def func(a=[], b={}):
print(a)
print(b)
print('#' * 12)
a.append(len(a)) # this will affect a's default value
b[len(a)] = len(a) # and this will affect b's one
func()
func()
func()
The parameters both have mutable default values. This means that, if you affect those objects, any modification will stick around in subsequent function calls. See if you can understand the output of those calls:
$ python arguments.defaults.mutable.py [] {} ############ [0] {1: 1} ############ [0, 1] {1: 1, 2: 2} ############
It's interesting, isn't it? While this behavior may seem very weird at first, it actually makes sense, and it's very handy, for example, when using memoization techniques (Google an example of that, if you're interested).
Even more interesting is what happens when, between the calls, we introduce one that doesn't use defaults, like this:
arguments.defaults.mutable.intermediate.call.py
func() func(a=[1, 2, 3], b={'B': 1}) func()
When we run this code, this is the output:
$ python arguments.defaults.mutable.intermediate.call.py [] {} ############ [1, 2, 3] {'B': 1} ############ [0] {1: 1} ############
This output shows us that the defaults are retained even if we call the function with other values. One question that comes to mind is, how do I get a fresh empty value every time? Well, the convention is the following:
arguments.defaults.mutable.no.trap.py
def func(a=None): if a is None: a = [] # do whatever you want with `a` ...
Note that, by using the preceding technique, if a
isn't passed when calling the function, you always get a brand new empty list.
Okay, enough with the input, let's look at the other side of the coin, the output.
Argument passing
Take a look at the following code. We declare a name x
in the global scope, then we declare a function func(y)
and we call it, passing x
. I highlighted the call in the code.
key.points.argument.passing.py
x = 3
def func(y):
print(y)
func(x) # prints: 3
When func
is called with x
, what happens is that within its local scope, a name y
is created, and it's pointed to the same object x
is pointing to. This is better clarified by the following picture:

The right part of the preceding picture depicts the state of the program when execution has reached the end, after func
has returned (None
). Take a look at the Frames column, and note that we have two names, x and func, in the global namespace (Global frame), pointing to an int (with a value of three) and to a function object, respectively. Right below it, in the rectangle titled func, we can see the function's local namespace, in which only one name has been defined: y. Because we have called func with x (line 5 in the left part of the picture), y is pointing to the same object that x is pointing to. This is what happens under the hood when an argument is passed to a function. If we had used the name x instead of y in the function definition, things would have been exactly the same (only maybe a bit confusing at first), there would be a local x in the function, and a global x outside, as we saw in the Scopes and name resolution section.
So, in a nutshell, what really happens is that the function creates in its local scope the names defined as arguments and, when we call it, we basically tell Python which objects those names must be pointed towards.
This is something that can be tricky to understand at first, so let's look at an example.
key.points.assignment.py
x = 3
def func(x):
x = 7 # defining a local x, not changing the global one
func(x)
print(x) # prints: 3
In the preceding code, when the line x = 7
is executed, what happens is that within the local scope of the function func
, the name x
is pointed to an integer with value 7, leaving the global x
unaltered.
This is the final point, and it's very important because Python apparently behaves differently with mutables (just apparently though). Let's look at an example:
key.points.mutable.py
x = [1, 2, 3]
def func(x):
x[1] = 42 # this affects the caller!
func(x)
print(x) # prints: [1, 42, 3]
Wow, we actually changed the original object! If you think about it, there is nothing weird in this behavior. The name x
in the function is set to point to the caller object by the function call and within the body of the function, we're not changing x
, in that we're not changing its reference, or, in other words, we are not changing the object x
is pointing to. What we're doing is accessing that object's element at position 1, and changing its value.
Remember point #2: "Assigning an object to an argument name within a function doesn't affect the caller". If that is clear to you, the following code should not be surprising.
key.points.mutable.assignment.py
x = [1, 2, 3] def func(x): x[1] = 42 # this changes the caller! x = 'something else' # this points x to a new string object func(x) print(x) # still prints: [1, 42, 3]
Take a look at the two lines I have highlighted. At first, we just access the caller object again, at position 1, and change its value to number 42. Then, we reassign x
to point to the string 'something else'
. This leaves the caller unaltered, according to point #2, and, in fact, the output is the same as that of the previous snippet.
Take your time to play around with this concept and experiment with prints and calls to the id
function until everything is clear in your mind. This is one of the key aspects of Python and it must be very clear, otherwise you risk introducing subtle bugs into your code.
Now that we have a good understanding of input parameters and how they behave, let's see how we can specify them.
There are five different ways of specifying input parameters. Let's look at them one by one.
Positional arguments are read from left to right and they are the most common type of arguments.
arguments.positional.py
def func(a, b, c): print(a, b, c) func(1, 2, 3) # prints: 1 2 3
There is not much else to say. They can be as numerous as you want and they are assigned by position. In the function call, 1
comes first, 2
comes second and 3
comes third, therefore they are assigned to a
, b
and c
respectively.
Keyword arguments are assigned by keyword using the name=value
syntax.
arguments.keyword.py
def func(a, b, c):
print(a, b, c)
func(a=1, c=2, b=3) # prints: 1 3 2
Keyword arguments act when calling the function instead of respecting the left-to-right positional assignment, k. Keyword arguments are matched by name, even when they don't respect the definition's original position (we'll see that there is a limitation to this behavior later, when we mix and match different types of arguments).
The counterpart of keyword arguments, on the definition side, is default values. The syntax is the same, name=value
, and allows us to not have to provide an argument if we are happy with the given default.
arguments.default.py
def func(a, b=4, c=88):
print(a, b, c)
func(1) # prints: 1 4 88
func(b=5, a=7, c=9) # prints: 7 5 9
func(42, c=9) # prints: 42 4 9
The are two things to notice, which are very important. First of all, you cannot specify a default argument on the left of a positional one. Second, note how in the examples, when an argument is passed without using the argument_name=value
syntax, it must be the first one in the list,, and it is always assigned to a
. Try and scramble those arguments and see what happens. Python error messages are very good at telling you what's wrong. So, for example, if you tried something like this:
func(b=1, c=2, 42) # positional argument after keyword one
You would get the following error:
SyntaxError: non-keyword arg after keyword arg
This informs you that you've called the function incorrectly.
Sometimes you may want to pass a variable number of positional arguments to a function and Python provides you with the ability to do it. Let's look at a very common use case, the minimum
function. This is a function that calculates the minimum of its input values.
arguments.variable.positional.py
def minimum(*n):
# print(n) # n is a tuple
if n: # explained after the code
mn = n[0]
for value in n[1:]:
if value < mn:
mn = value
print(mn)
minimum(1, 3, -7, 9) # n = (1, 3, -7, 9) - prints: -7
minimum() # n = () - prints: nothing
As you can see, when we specify a parameter prepending a *
to its name, we are telling Python that that parameter will be collecting a variable number of positional arguments, according to how the function is called. Within the function, n
is a tuple. Uncomment the print(n)
to see for yourself and play around with it for a bit.
Note
Have you noticed how we checked if n
wasn't empty with a simple if n:
? This is due to the fact that collection objects evaluate to True
when non-empty, and otherwise False
in Python. This is true for tuples, sets, lists, dictionaries, and so on.
One other thing to note is that we may want to throw an error when we call the function with no arguments, instead of silently doing nothing. In this context, we're not concerned about making this function robust, but in understanding variable positional arguments.
Let's make another example to show you two things that, in my experience, are confusing to those who are new to this.
arguments.variable.positional.unpacking.py
def func(*args): print(args) values = (1, 3, -7, 9) func(values) # equivalent to: func((1, 3, -7, 9)) func(*values) # equivalent to: func(1, 3, -7, 9)
Take a good look at the last two lines of the preceding example. In the first one, we call func
with one argument, a four elements tuple. In the second example, by using the *
syntax, we're doing something called unpacking, which means that the four elements tuple is unpacked, and the function is called with four arguments: 1, 3, -7, 9
.
This behavior is part of the magic Python does to allow you to do amazing things when calling functions dynamically.
Variable keyword arguments are very similar to variable positional arguments. The only difference is the syntax (**
instead of *
) and that they are collected in a dictionary. Collection and unpacking work in the same way, so let's look at an example:
arguments.variable.keyword.py
def func(**kwargs): print(kwargs) # All calls equivalent. They print: {'a': 1, 'b': 42} func(a=1, b=42) func(**{'a': 1, 'b': 42}) func(**dict(a=1, b=42))
All the calls are equivalent in the preceding example. You can see that adding a **
in front of the parameter name in the function definition tells Python to use that name to collect a variable number of keyword parameters. On the other hand, when we call the function, we can either pass name=value
arguments explicitly, or unpack a dictionary using the same **
syntax.
The reason why being able to pass a variable number of keyword parameters is so important may not be evident at the moment, so, how about a more realistic example? Let's define a function that connects to a database. We want to connect to a default database by simply calling this function with no parameters. We also want to connect to any other database by passing the function the appropriate arguments. Before you read on, spend a couple of minutes figuring out a solution by yourself.
arguments.variable.db.py
def connect(**options): conn_params = { 'host': options.get('host', '127.0.0.1'), 'port': options.get('port', 5432), 'user': options.get('user', ''), 'pwd': options.get('pwd', ''), } print(conn_params) # we then connect to the db (commented out) # db.connect(**conn_params) connect() connect(host='127.0.0.42', port=5433) connect(port=5431, user='fab', pwd='gandalf')
Note in the function we can prepare a dictionary of connection parameters (conn_params
) in the function using default values as fallback, allowing them to be overwritten if they are provided in the function call. There are better ways to do this with fewer lines of code but we're not concerned with that now. Running the preceding code yields the following result:
$ python arguments.variable.db.py {'host': '127.0.0.1', 'pwd': '', 'user': '', 'port': 5432} {'host': '127.0.0.42', 'pwd': '', 'user': '', 'port': 5433} {'host': '127.0.0.1', 'pwd': 'gandalf', 'user': 'fab', 'port': 5431}
Note the correspondence between the function calls and the output. Note how default values are either there or overridden, according to what was passed to the function.
Python 3 allows for a new type of parameter: the keyword-only parameter. We are going to study them only briefly as their use cases are not that frequent. There are two ways of specifying them, either after the variable positional arguments, or after a bare *. Let's see an example of both.
arguments.keyword.only.py
def kwo(*a, c): print(a, c) kwo(1, 2, 3, c=7) # prints: (1, 2, 3) 7 kwo(c=4) # prints: () 4 # kwo(1, 2) # breaks, invalid syntax, with the following error # TypeError: kwo() missing 1 required keyword-only argument: 'c' def kwo2(a, b=42, *, c): print(a, b, c) kwo2(3, b=7, c=99) # prints: 3 7 99 kwo2(3, c=13) # prints: 3 42 13 # kwo2(3, 23) # breaks, invalid syntax, with the following error # TypeError: kwo2() missing 1 required keyword-only argument: 'c'
As anticipated, the function, kwo
, takes a variable number of positional arguments (a
) and a keyword-only function, c
. The results of the calls are straightforward and you can uncomment the third call to see what error Python returns.
The same applies to the function, kwo2
, which differs from kwo
in that it takes a positional argument a
, a keyword argument b
, and then a keyword-only argument, c
. You can uncomment the third call to see the error.
Now that you know how to specify different types of input parameters, let's see how you can combine them in function definitions.
You can combine input parameters, as long as you follow these ordering rules:
- When defining a function, normal positional arguments come first (
name
), then any default arguments (name=value
), then the variable positional arguments (*name
, or simply*
), then any keyword-only arguments (eithername
orname=value
form is good), then any variable keyword arguments (**name
). - On the other hand, when calling a function, arguments must be given in the following order: positional arguments first (
value
), then any combination of keyword arguments (name=value
), variable positional arguments (*name
), then variable keyword arguments (**name
).
Since this can be a bit tricky when left hanging in the theoretical world, let's look at a couple of quick examples.
arguments.all.py
def func(a, b, c=7, *args, **kwargs): print('a, b, c:', a, b, c) print('args:', args) print('kwargs:', kwargs) func(1, 2, 3, *(5, 7, 9), **{'A': 'a', 'B': 'b'}) func(1, 2, 3, 5, 7, 9, A='a', B='b') # same as previous one
Note the order of the parameters in the function definition, and that the two calls are equivalent. In the first one, we're using the unpacking operators for iterables and dictionaries, while in the second one we're using a more explicit syntax. The execution of this yields (I printed only the result of one call):
$ python arguments.all.py a, b, c: 1 2 3 args: (5, 7, 9) kwargs: {'A': 'a', 'B': 'b'}
Let's now look at an example with keyword-only arguments.
arguments.all.kwonly.py
def func_with_kwonly(a, b=42, *args, c, d=256, **kwargs):
print('a, b:', a, b)
print('c, d:', c, d)
print('args:', args)
print('kwargs:', kwargs)
# both calls equivalent
func_with_kwonly(3, 42, c=0, d=1, *(7, 9, 11), e='E', f='F')
func_with_kwonly(3, 42, *(7, 9, 11), c=0, d=1, e='E', f='F')
Note that I have highlighted the keyword-only arguments in the function declaration. They come after the variable positional argument *args
, and it would be the same if they came right after a single *
(in which case there wouldn't be a variable positional argument). The execution of this yields (I printed only the result of one call):
$ python arguments.all.kwonly.py a, b: 3 42 c, d: 0 1 args: (7, 9, 11) kwargs: {'f': 'F', 'e': 'E'}
One other thing to note are the names I gave to the variable positional and keyword arguments. You're free to choose differently, but be aware that args
and kwargs
are the conventional names given to these parameters, at least generically. Now that you know how to define a function in all possible flavors, let me show you something tricky: mutable defaults.
One thing to be very aware of with Python is that default values are created at def
time, therefore, subsequent calls to the same function will possibly behave differently according to the mutability of their default values. Let's look at an example:
arguments.defaults.mutable.py
def func(a=[], b={}):
print(a)
print(b)
print('#' * 12)
a.append(len(a)) # this will affect a's default value
b[len(a)] = len(a) # and this will affect b's one
func()
func()
func()
The parameters both have mutable default values. This means that, if you affect those objects, any modification will stick around in subsequent function calls. See if you can understand the output of those calls:
$ python arguments.defaults.mutable.py [] {} ############ [0] {1: 1} ############ [0, 1] {1: 1, 2: 2} ############
It's interesting, isn't it? While this behavior may seem very weird at first, it actually makes sense, and it's very handy, for example, when using memoization techniques (Google an example of that, if you're interested).
Even more interesting is what happens when, between the calls, we introduce one that doesn't use defaults, like this:
arguments.defaults.mutable.intermediate.call.py
func() func(a=[1, 2, 3], b={'B': 1}) func()
When we run this code, this is the output:
$ python arguments.defaults.mutable.intermediate.call.py [] {} ############ [1, 2, 3] {'B': 1} ############ [0] {1: 1} ############
This output shows us that the defaults are retained even if we call the function with other values. One question that comes to mind is, how do I get a fresh empty value every time? Well, the convention is the following:
arguments.defaults.mutable.no.trap.py
def func(a=None): if a is None: a = [] # do whatever you want with `a` ...
Note that, by using the preceding technique, if a
isn't passed when calling the function, you always get a brand new empty list.
Okay, enough with the input, let's look at the other side of the coin, the output.
Assignment to argument names don't affect the caller
This is something that can be tricky to understand at first, so let's look at an example.
key.points.assignment.py
x = 3
def func(x):
x = 7 # defining a local x, not changing the global one
func(x)
print(x) # prints: 3
In the preceding code, when the line x = 7
is executed, what happens is that within the local scope of the function func
, the name x
is pointed to an integer with value 7, leaving the global x
unaltered.
This is the final point, and it's very important because Python apparently behaves differently with mutables (just apparently though). Let's look at an example:
key.points.mutable.py
x = [1, 2, 3]
def func(x):
x[1] = 42 # this affects the caller!
func(x)
print(x) # prints: [1, 42, 3]
Wow, we actually changed the original object! If you think about it, there is nothing weird in this behavior. The name x
in the function is set to point to the caller object by the function call and within the body of the function, we're not changing x
, in that we're not changing its reference, or, in other words, we are not changing the object x
is pointing to. What we're doing is accessing that object's element at position 1, and changing its value.
Remember point #2: "Assigning an object to an argument name within a function doesn't affect the caller". If that is clear to you, the following code should not be surprising.
key.points.mutable.assignment.py
x = [1, 2, 3] def func(x): x[1] = 42 # this changes the caller! x = 'something else' # this points x to a new string object func(x) print(x) # still prints: [1, 42, 3]
Take a look at the two lines I have highlighted. At first, we just access the caller object again, at position 1, and change its value to number 42. Then, we reassign x
to point to the string 'something else'
. This leaves the caller unaltered, according to point #2, and, in fact, the output is the same as that of the previous snippet.
Take your time to play around with this concept and experiment with prints and calls to the id
function until everything is clear in your mind. This is one of the key aspects of Python and it must be very clear, otherwise you risk introducing subtle bugs into your code.
Now that we have a good understanding of input parameters and how they behave, let's see how we can specify them.
There are five different ways of specifying input parameters. Let's look at them one by one.
Positional arguments are read from left to right and they are the most common type of arguments.
arguments.positional.py
def func(a, b, c): print(a, b, c) func(1, 2, 3) # prints: 1 2 3
There is not much else to say. They can be as numerous as you want and they are assigned by position. In the function call, 1
comes first, 2
comes second and 3
comes third, therefore they are assigned to a
, b
and c
respectively.
Keyword arguments are assigned by keyword using the name=value
syntax.
arguments.keyword.py
def func(a, b, c):
print(a, b, c)
func(a=1, c=2, b=3) # prints: 1 3 2
Keyword arguments act when calling the function instead of respecting the left-to-right positional assignment, k. Keyword arguments are matched by name, even when they don't respect the definition's original position (we'll see that there is a limitation to this behavior later, when we mix and match different types of arguments).
The counterpart of keyword arguments, on the definition side, is default values. The syntax is the same, name=value
, and allows us to not have to provide an argument if we are happy with the given default.
arguments.default.py
def func(a, b=4, c=88):
print(a, b, c)
func(1) # prints: 1 4 88
func(b=5, a=7, c=9) # prints: 7 5 9
func(42, c=9) # prints: 42 4 9
The are two things to notice, which are very important. First of all, you cannot specify a default argument on the left of a positional one. Second, note how in the examples, when an argument is passed without using the argument_name=value
syntax, it must be the first one in the list,, and it is always assigned to a
. Try and scramble those arguments and see what happens. Python error messages are very good at telling you what's wrong. So, for example, if you tried something like this:
func(b=1, c=2, 42) # positional argument after keyword one
You would get the following error:
SyntaxError: non-keyword arg after keyword arg
This informs you that you've called the function incorrectly.
Sometimes you may want to pass a variable number of positional arguments to a function and Python provides you with the ability to do it. Let's look at a very common use case, the minimum
function. This is a function that calculates the minimum of its input values.
arguments.variable.positional.py
def minimum(*n):
# print(n) # n is a tuple
if n: # explained after the code
mn = n[0]
for value in n[1:]:
if value < mn:
mn = value
print(mn)
minimum(1, 3, -7, 9) # n = (1, 3, -7, 9) - prints: -7
minimum() # n = () - prints: nothing
As you can see, when we specify a parameter prepending a *
to its name, we are telling Python that that parameter will be collecting a variable number of positional arguments, according to how the function is called. Within the function, n
is a tuple. Uncomment the print(n)
to see for yourself and play around with it for a bit.
Note
Have you noticed how we checked if n
wasn't empty with a simple if n:
? This is due to the fact that collection objects evaluate to True
when non-empty, and otherwise False
in Python. This is true for tuples, sets, lists, dictionaries, and so on.
One other thing to note is that we may want to throw an error when we call the function with no arguments, instead of silently doing nothing. In this context, we're not concerned about making this function robust, but in understanding variable positional arguments.
Let's make another example to show you two things that, in my experience, are confusing to those who are new to this.
arguments.variable.positional.unpacking.py
def func(*args): print(args) values = (1, 3, -7, 9) func(values) # equivalent to: func((1, 3, -7, 9)) func(*values) # equivalent to: func(1, 3, -7, 9)
Take a good look at the last two lines of the preceding example. In the first one, we call func
with one argument, a four elements tuple. In the second example, by using the *
syntax, we're doing something called unpacking, which means that the four elements tuple is unpacked, and the function is called with four arguments: 1, 3, -7, 9
.
This behavior is part of the magic Python does to allow you to do amazing things when calling functions dynamically.
Variable keyword arguments are very similar to variable positional arguments. The only difference is the syntax (**
instead of *
) and that they are collected in a dictionary. Collection and unpacking work in the same way, so let's look at an example:
arguments.variable.keyword.py
def func(**kwargs): print(kwargs) # All calls equivalent. They print: {'a': 1, 'b': 42} func(a=1, b=42) func(**{'a': 1, 'b': 42}) func(**dict(a=1, b=42))
All the calls are equivalent in the preceding example. You can see that adding a **
in front of the parameter name in the function definition tells Python to use that name to collect a variable number of keyword parameters. On the other hand, when we call the function, we can either pass name=value
arguments explicitly, or unpack a dictionary using the same **
syntax.
The reason why being able to pass a variable number of keyword parameters is so important may not be evident at the moment, so, how about a more realistic example? Let's define a function that connects to a database. We want to connect to a default database by simply calling this function with no parameters. We also want to connect to any other database by passing the function the appropriate arguments. Before you read on, spend a couple of minutes figuring out a solution by yourself.
arguments.variable.db.py
def connect(**options): conn_params = { 'host': options.get('host', '127.0.0.1'), 'port': options.get('port', 5432), 'user': options.get('user', ''), 'pwd': options.get('pwd', ''), } print(conn_params) # we then connect to the db (commented out) # db.connect(**conn_params) connect() connect(host='127.0.0.42', port=5433) connect(port=5431, user='fab', pwd='gandalf')
Note in the function we can prepare a dictionary of connection parameters (conn_params
) in the function using default values as fallback, allowing them to be overwritten if they are provided in the function call. There are better ways to do this with fewer lines of code but we're not concerned with that now. Running the preceding code yields the following result:
$ python arguments.variable.db.py {'host': '127.0.0.1', 'pwd': '', 'user': '', 'port': 5432} {'host': '127.0.0.42', 'pwd': '', 'user': '', 'port': 5433} {'host': '127.0.0.1', 'pwd': 'gandalf', 'user': 'fab', 'port': 5431}
Note the correspondence between the function calls and the output. Note how default values are either there or overridden, according to what was passed to the function.
Python 3 allows for a new type of parameter: the keyword-only parameter. We are going to study them only briefly as their use cases are not that frequent. There are two ways of specifying them, either after the variable positional arguments, or after a bare *. Let's see an example of both.
arguments.keyword.only.py
def kwo(*a, c): print(a, c) kwo(1, 2, 3, c=7) # prints: (1, 2, 3) 7 kwo(c=4) # prints: () 4 # kwo(1, 2) # breaks, invalid syntax, with the following error # TypeError: kwo() missing 1 required keyword-only argument: 'c' def kwo2(a, b=42, *, c): print(a, b, c) kwo2(3, b=7, c=99) # prints: 3 7 99 kwo2(3, c=13) # prints: 3 42 13 # kwo2(3, 23) # breaks, invalid syntax, with the following error # TypeError: kwo2() missing 1 required keyword-only argument: 'c'
As anticipated, the function, kwo
, takes a variable number of positional arguments (a
) and a keyword-only function, c
. The results of the calls are straightforward and you can uncomment the third call to see what error Python returns.
The same applies to the function, kwo2
, which differs from kwo
in that it takes a positional argument a
, a keyword argument b
, and then a keyword-only argument, c
. You can uncomment the third call to see the error.
Now that you know how to specify different types of input parameters, let's see how you can combine them in function definitions.
You can combine input parameters, as long as you follow these ordering rules:
- When defining a function, normal positional arguments come first (
name
), then any default arguments (name=value
), then the variable positional arguments (*name
, or simply*
), then any keyword-only arguments (eithername
orname=value
form is good), then any variable keyword arguments (**name
). - On the other hand, when calling a function, arguments must be given in the following order: positional arguments first (
value
), then any combination of keyword arguments (name=value
), variable positional arguments (*name
), then variable keyword arguments (**name
).
Since this can be a bit tricky when left hanging in the theoretical world, let's look at a couple of quick examples.
arguments.all.py
def func(a, b, c=7, *args, **kwargs): print('a, b, c:', a, b, c) print('args:', args) print('kwargs:', kwargs) func(1, 2, 3, *(5, 7, 9), **{'A': 'a', 'B': 'b'}) func(1, 2, 3, 5, 7, 9, A='a', B='b') # same as previous one
Note the order of the parameters in the function definition, and that the two calls are equivalent. In the first one, we're using the unpacking operators for iterables and dictionaries, while in the second one we're using a more explicit syntax. The execution of this yields (I printed only the result of one call):
$ python arguments.all.py a, b, c: 1 2 3 args: (5, 7, 9) kwargs: {'A': 'a', 'B': 'b'}
Let's now look at an example with keyword-only arguments.
arguments.all.kwonly.py
def func_with_kwonly(a, b=42, *args, c, d=256, **kwargs):
print('a, b:', a, b)
print('c, d:', c, d)
print('args:', args)
print('kwargs:', kwargs)
# both calls equivalent
func_with_kwonly(3, 42, c=0, d=1, *(7, 9, 11), e='E', f='F')
func_with_kwonly(3, 42, *(7, 9, 11), c=0, d=1, e='E', f='F')
Note that I have highlighted the keyword-only arguments in the function declaration. They come after the variable positional argument *args
, and it would be the same if they came right after a single *
(in which case there wouldn't be a variable positional argument). The execution of this yields (I printed only the result of one call):
$ python arguments.all.kwonly.py a, b: 3 42 c, d: 0 1 args: (7, 9, 11) kwargs: {'f': 'F', 'e': 'E'}
One other thing to note are the names I gave to the variable positional and keyword arguments. You're free to choose differently, but be aware that args
and kwargs
are the conventional names given to these parameters, at least generically. Now that you know how to define a function in all possible flavors, let me show you something tricky: mutable defaults.
One thing to be very aware of with Python is that default values are created at def
time, therefore, subsequent calls to the same function will possibly behave differently according to the mutability of their default values. Let's look at an example:
arguments.defaults.mutable.py
def func(a=[], b={}):
print(a)
print(b)
print('#' * 12)
a.append(len(a)) # this will affect a's default value
b[len(a)] = len(a) # and this will affect b's one
func()
func()
func()
The parameters both have mutable default values. This means that, if you affect those objects, any modification will stick around in subsequent function calls. See if you can understand the output of those calls:
$ python arguments.defaults.mutable.py [] {} ############ [0] {1: 1} ############ [0, 1] {1: 1, 2: 2} ############
It's interesting, isn't it? While this behavior may seem very weird at first, it actually makes sense, and it's very handy, for example, when using memoization techniques (Google an example of that, if you're interested).
Even more interesting is what happens when, between the calls, we introduce one that doesn't use defaults, like this:
arguments.defaults.mutable.intermediate.call.py
func() func(a=[1, 2, 3], b={'B': 1}) func()
When we run this code, this is the output:
$ python arguments.defaults.mutable.intermediate.call.py [] {} ############ [1, 2, 3] {'B': 1} ############ [0] {1: 1} ############
This output shows us that the defaults are retained even if we call the function with other values. One question that comes to mind is, how do I get a fresh empty value every time? Well, the convention is the following:
arguments.defaults.mutable.no.trap.py
def func(a=None): if a is None: a = [] # do whatever you want with `a` ...
Note that, by using the preceding technique, if a
isn't passed when calling the function, you always get a brand new empty list.
Okay, enough with the input, let's look at the other side of the coin, the output.
Changing a mutable affects the caller
This is the final point, and it's very important because Python apparently behaves differently with mutables (just apparently though). Let's look at an example:
key.points.mutable.py
x = [1, 2, 3]
def func(x):
x[1] = 42 # this affects the caller!
func(x)
print(x) # prints: [1, 42, 3]
Wow, we actually changed the original object! If you think about it, there is nothing weird in this behavior. The name x
in the function is set to point to the caller object by the function call and within the body of the function, we're not changing x
, in that we're not changing its reference, or, in other words, we are not changing the object x
is pointing to. What we're doing is accessing that object's element at position 1, and changing its value.
Remember point #2: "Assigning an object to an argument name within a function doesn't affect the caller". If that is clear to you, the following code should not be surprising.
key.points.mutable.assignment.py
x = [1, 2, 3] def func(x): x[1] = 42 # this changes the caller! x = 'something else' # this points x to a new string object func(x) print(x) # still prints: [1, 42, 3]
Take a look at the two lines I have highlighted. At first, we just access the caller object again, at position 1, and change its value to number 42. Then, we reassign x
to point to the string 'something else'
. This leaves the caller unaltered, according to point #2, and, in fact, the output is the same as that of the previous snippet.
Take your time to play around with this concept and experiment with prints and calls to the id
function until everything is clear in your mind. This is one of the key aspects of Python and it must be very clear, otherwise you risk introducing subtle bugs into your code.
Now that we have a good understanding of input parameters and how they behave, let's see how we can specify them.
There are five different ways of specifying input parameters. Let's look at them one by one.
Positional arguments are read from left to right and they are the most common type of arguments.
arguments.positional.py
def func(a, b, c): print(a, b, c) func(1, 2, 3) # prints: 1 2 3
There is not much else to say. They can be as numerous as you want and they are assigned by position. In the function call, 1
comes first, 2
comes second and 3
comes third, therefore they are assigned to a
, b
and c
respectively.
Keyword arguments are assigned by keyword using the name=value
syntax.
arguments.keyword.py
def func(a, b, c):
print(a, b, c)
func(a=1, c=2, b=3) # prints: 1 3 2
Keyword arguments act when calling the function instead of respecting the left-to-right positional assignment, k. Keyword arguments are matched by name, even when they don't respect the definition's original position (we'll see that there is a limitation to this behavior later, when we mix and match different types of arguments).
The counterpart of keyword arguments, on the definition side, is default values. The syntax is the same, name=value
, and allows us to not have to provide an argument if we are happy with the given default.
arguments.default.py
def func(a, b=4, c=88):
print(a, b, c)
func(1) # prints: 1 4 88
func(b=5, a=7, c=9) # prints: 7 5 9
func(42, c=9) # prints: 42 4 9
The are two things to notice, which are very important. First of all, you cannot specify a default argument on the left of a positional one. Second, note how in the examples, when an argument is passed without using the argument_name=value
syntax, it must be the first one in the list,, and it is always assigned to a
. Try and scramble those arguments and see what happens. Python error messages are very good at telling you what's wrong. So, for example, if you tried something like this:
func(b=1, c=2, 42) # positional argument after keyword one
You would get the following error:
SyntaxError: non-keyword arg after keyword arg
This informs you that you've called the function incorrectly.
Sometimes you may want to pass a variable number of positional arguments to a function and Python provides you with the ability to do it. Let's look at a very common use case, the minimum
function. This is a function that calculates the minimum of its input values.
arguments.variable.positional.py
def minimum(*n):
# print(n) # n is a tuple
if n: # explained after the code
mn = n[0]
for value in n[1:]:
if value < mn:
mn = value
print(mn)
minimum(1, 3, -7, 9) # n = (1, 3, -7, 9) - prints: -7
minimum() # n = () - prints: nothing
As you can see, when we specify a parameter prepending a *
to its name, we are telling Python that that parameter will be collecting a variable number of positional arguments, according to how the function is called. Within the function, n
is a tuple. Uncomment the print(n)
to see for yourself and play around with it for a bit.
Note
Have you noticed how we checked if n
wasn't empty with a simple if n:
? This is due to the fact that collection objects evaluate to True
when non-empty, and otherwise False
in Python. This is true for tuples, sets, lists, dictionaries, and so on.
One other thing to note is that we may want to throw an error when we call the function with no arguments, instead of silently doing nothing. In this context, we're not concerned about making this function robust, but in understanding variable positional arguments.
Let's make another example to show you two things that, in my experience, are confusing to those who are new to this.
arguments.variable.positional.unpacking.py
def func(*args): print(args) values = (1, 3, -7, 9) func(values) # equivalent to: func((1, 3, -7, 9)) func(*values) # equivalent to: func(1, 3, -7, 9)
Take a good look at the last two lines of the preceding example. In the first one, we call func
with one argument, a four elements tuple. In the second example, by using the *
syntax, we're doing something called unpacking, which means that the four elements tuple is unpacked, and the function is called with four arguments: 1, 3, -7, 9
.
This behavior is part of the magic Python does to allow you to do amazing things when calling functions dynamically.
Variable keyword arguments are very similar to variable positional arguments. The only difference is the syntax (**
instead of *
) and that they are collected in a dictionary. Collection and unpacking work in the same way, so let's look at an example:
arguments.variable.keyword.py
def func(**kwargs): print(kwargs) # All calls equivalent. They print: {'a': 1, 'b': 42} func(a=1, b=42) func(**{'a': 1, 'b': 42}) func(**dict(a=1, b=42))
All the calls are equivalent in the preceding example. You can see that adding a **
in front of the parameter name in the function definition tells Python to use that name to collect a variable number of keyword parameters. On the other hand, when we call the function, we can either pass name=value
arguments explicitly, or unpack a dictionary using the same **
syntax.
The reason why being able to pass a variable number of keyword parameters is so important may not be evident at the moment, so, how about a more realistic example? Let's define a function that connects to a database. We want to connect to a default database by simply calling this function with no parameters. We also want to connect to any other database by passing the function the appropriate arguments. Before you read on, spend a couple of minutes figuring out a solution by yourself.
arguments.variable.db.py
def connect(**options): conn_params = { 'host': options.get('host', '127.0.0.1'), 'port': options.get('port', 5432), 'user': options.get('user', ''), 'pwd': options.get('pwd', ''), } print(conn_params) # we then connect to the db (commented out) # db.connect(**conn_params) connect() connect(host='127.0.0.42', port=5433) connect(port=5431, user='fab', pwd='gandalf')
Note in the function we can prepare a dictionary of connection parameters (conn_params
) in the function using default values as fallback, allowing them to be overwritten if they are provided in the function call. There are better ways to do this with fewer lines of code but we're not concerned with that now. Running the preceding code yields the following result:
$ python arguments.variable.db.py {'host': '127.0.0.1', 'pwd': '', 'user': '', 'port': 5432} {'host': '127.0.0.42', 'pwd': '', 'user': '', 'port': 5433} {'host': '127.0.0.1', 'pwd': 'gandalf', 'user': 'fab', 'port': 5431}
Note the correspondence between the function calls and the output. Note how default values are either there or overridden, according to what was passed to the function.
Python 3 allows for a new type of parameter: the keyword-only parameter. We are going to study them only briefly as their use cases are not that frequent. There are two ways of specifying them, either after the variable positional arguments, or after a bare *. Let's see an example of both.
arguments.keyword.only.py
def kwo(*a, c): print(a, c) kwo(1, 2, 3, c=7) # prints: (1, 2, 3) 7 kwo(c=4) # prints: () 4 # kwo(1, 2) # breaks, invalid syntax, with the following error # TypeError: kwo() missing 1 required keyword-only argument: 'c' def kwo2(a, b=42, *, c): print(a, b, c) kwo2(3, b=7, c=99) # prints: 3 7 99 kwo2(3, c=13) # prints: 3 42 13 # kwo2(3, 23) # breaks, invalid syntax, with the following error # TypeError: kwo2() missing 1 required keyword-only argument: 'c'
As anticipated, the function, kwo
, takes a variable number of positional arguments (a
) and a keyword-only function, c
. The results of the calls are straightforward and you can uncomment the third call to see what error Python returns.
The same applies to the function, kwo2
, which differs from kwo
in that it takes a positional argument a
, a keyword argument b
, and then a keyword-only argument, c
. You can uncomment the third call to see the error.
Now that you know how to specify different types of input parameters, let's see how you can combine them in function definitions.
You can combine input parameters, as long as you follow these ordering rules:
- When defining a function, normal positional arguments come first (
name
), then any default arguments (name=value
), then the variable positional arguments (*name
, or simply*
), then any keyword-only arguments (eithername
orname=value
form is good), then any variable keyword arguments (**name
). - On the other hand, when calling a function, arguments must be given in the following order: positional arguments first (
value
), then any combination of keyword arguments (name=value
), variable positional arguments (*name
), then variable keyword arguments (**name
).
Since this can be a bit tricky when left hanging in the theoretical world, let's look at a couple of quick examples.
arguments.all.py
def func(a, b, c=7, *args, **kwargs): print('a, b, c:', a, b, c) print('args:', args) print('kwargs:', kwargs) func(1, 2, 3, *(5, 7, 9), **{'A': 'a', 'B': 'b'}) func(1, 2, 3, 5, 7, 9, A='a', B='b') # same as previous one
Note the order of the parameters in the function definition, and that the two calls are equivalent. In the first one, we're using the unpacking operators for iterables and dictionaries, while in the second one we're using a more explicit syntax. The execution of this yields (I printed only the result of one call):
$ python arguments.all.py a, b, c: 1 2 3 args: (5, 7, 9) kwargs: {'A': 'a', 'B': 'b'}
Let's now look at an example with keyword-only arguments.
arguments.all.kwonly.py
def func_with_kwonly(a, b=42, *args, c, d=256, **kwargs):
print('a, b:', a, b)
print('c, d:', c, d)
print('args:', args)
print('kwargs:', kwargs)
# both calls equivalent
func_with_kwonly(3, 42, c=0, d=1, *(7, 9, 11), e='E', f='F')
func_with_kwonly(3, 42, *(7, 9, 11), c=0, d=1, e='E', f='F')
Note that I have highlighted the keyword-only arguments in the function declaration. They come after the variable positional argument *args
, and it would be the same if they came right after a single *
(in which case there wouldn't be a variable positional argument). The execution of this yields (I printed only the result of one call):
$ python arguments.all.kwonly.py a, b: 3 42 c, d: 0 1 args: (7, 9, 11) kwargs: {'f': 'F', 'e': 'E'}
One other thing to note are the names I gave to the variable positional and keyword arguments. You're free to choose differently, but be aware that args
and kwargs
are the conventional names given to these parameters, at least generically. Now that you know how to define a function in all possible flavors, let me show you something tricky: mutable defaults.
One thing to be very aware of with Python is that default values are created at def
time, therefore, subsequent calls to the same function will possibly behave differently according to the mutability of their default values. Let's look at an example:
arguments.defaults.mutable.py
def func(a=[], b={}):
print(a)
print(b)
print('#' * 12)
a.append(len(a)) # this will affect a's default value
b[len(a)] = len(a) # and this will affect b's one
func()
func()
func()
The parameters both have mutable default values. This means that, if you affect those objects, any modification will stick around in subsequent function calls. See if you can understand the output of those calls:
$ python arguments.defaults.mutable.py [] {} ############ [0] {1: 1} ############ [0, 1] {1: 1, 2: 2} ############
It's interesting, isn't it? While this behavior may seem very weird at first, it actually makes sense, and it's very handy, for example, when using memoization techniques (Google an example of that, if you're interested).
Even more interesting is what happens when, between the calls, we introduce one that doesn't use defaults, like this:
arguments.defaults.mutable.intermediate.call.py
func() func(a=[1, 2, 3], b={'B': 1}) func()
When we run this code, this is the output:
$ python arguments.defaults.mutable.intermediate.call.py [] {} ############ [1, 2, 3] {'B': 1} ############ [0] {1: 1} ############
This output shows us that the defaults are retained even if we call the function with other values. One question that comes to mind is, how do I get a fresh empty value every time? Well, the convention is the following:
arguments.defaults.mutable.no.trap.py
def func(a=None): if a is None: a = [] # do whatever you want with `a` ...
Note that, by using the preceding technique, if a
isn't passed when calling the function, you always get a brand new empty list.
Okay, enough with the input, let's look at the other side of the coin, the output.
How to specify input parameters
There are five different ways of specifying input parameters. Let's look at them one by one.
Positional arguments are read from left to right and they are the most common type of arguments.
arguments.positional.py
def func(a, b, c): print(a, b, c) func(1, 2, 3) # prints: 1 2 3
There is not much else to say. They can be as numerous as you want and they are assigned by position. In the function call, 1
comes first, 2
comes second and 3
comes third, therefore they are assigned to a
, b
and c
respectively.
Keyword arguments are assigned by keyword using the name=value
syntax.
arguments.keyword.py
def func(a, b, c):
print(a, b, c)
func(a=1, c=2, b=3) # prints: 1 3 2
Keyword arguments act when calling the function instead of respecting the left-to-right positional assignment, k. Keyword arguments are matched by name, even when they don't respect the definition's original position (we'll see that there is a limitation to this behavior later, when we mix and match different types of arguments).
The counterpart of keyword arguments, on the definition side, is default values. The syntax is the same, name=value
, and allows us to not have to provide an argument if we are happy with the given default.
arguments.default.py
def func(a, b=4, c=88):
print(a, b, c)
func(1) # prints: 1 4 88
func(b=5, a=7, c=9) # prints: 7 5 9
func(42, c=9) # prints: 42 4 9
The are two things to notice, which are very important. First of all, you cannot specify a default argument on the left of a positional one. Second, note how in the examples, when an argument is passed without using the argument_name=value
syntax, it must be the first one in the list,, and it is always assigned to a
. Try and scramble those arguments and see what happens. Python error messages are very good at telling you what's wrong. So, for example, if you tried something like this:
func(b=1, c=2, 42) # positional argument after keyword one
You would get the following error:
SyntaxError: non-keyword arg after keyword arg
This informs you that you've called the function incorrectly.
Sometimes you may want to pass a variable number of positional arguments to a function and Python provides you with the ability to do it. Let's look at a very common use case, the minimum
function. This is a function that calculates the minimum of its input values.
arguments.variable.positional.py
def minimum(