This is the second post of my series about working around Google Drive APIs. If you haven’t read my first articleon Google Drive APIs, I urge you to do it. It gives some context that might be required for understanding everything I’m referring to. Not to mention that it might be useful for you on its own anyway. I touched some areas I think the Google Drive API lacks — like methods used by the UI not available publicly. I mentioned constraints present due to the original underlying architecture: how folders were acting like labels. Furthermore, I highlighted how some small issues in the Drive’s backend can have a huge influence on products using it — like noop actions triggering change notifications.
So… let’s start with the toughest nut to crack. I say this because by far this was the biggest black box and has cost us the most time to address in accordance with our quality standards. In many cases, we had to strive to achieve as high throughput as possible which resulted in multiple iterations to tweak/perfect the system. Despite how minor a new information available might have seemed it could have easily resulted in a refactor with a considerable size.
Oh boy, you are in a world of pain on this one. Do you know what is the official recommendation for mitigating the rate limit exceeded issues? Use exponential backoff. That’s it. You would probably think, they would help you in some way in avoiding spamming the API unnecessarily, right? Nah. That’s for losers. You are a pro. Solve it yourself. (Check how GitHub API handles rate limits for example, and you will understand what I’m talking about). There is no official explanation for how they calculate it, how it works, nothing. All you can do is do benchmarks and based on the findings do educate guesses. If you are lucky you might get hints from support. Apart from that…
Yes, exponential backoff works… and it’s enough… until a point. When you reach that and you see that so much of your resources — and essentially money — get burnt on requests that get rejected due to a rate limit you start to think how to optimize it.
The first step — to go under the rate limit by a big margin so you won’t have issues — could be rather easy.
The next one on the other hand — when you try to do as many requests as possible because you want lower latency and/or you just have a lot of tasks to do and you need the throughput — is where it gets rather tricky. You essentially have to implement your own solution that changes the throughput based on the rate limits you encounter. Don’t get me wrong. Using exponential backoff, and reducing the throughput on our side if we reach the rate limit is a “must-have” functionality. My problem is that’s it’s the ONLY input the whole system can use.
Then if this wouldn’t be enough the rate changes based on the nature of your calls. This is where the educated guesses come into play:
I thought about creating fancy charts for this, but I realized it might change by the time you read this. It has already changed multiple times over the years. But don’t be sad. I certainly don’t mind. In fact, I enjoy it. I like the challenge of doing the same task over and over again to investigate the behavior.
Was it convincing enough? Anyway, let’s just hope it’s still valid and use these relatively safe magic numbers. Sustained read 10/s, sustained write 1/s.
Ohh, almost forgot. The rate limiting is applied on a per-user basis, and a file belongs to the owner and not the actual user we do the modification with. So if you want to implement some throttling you will have to know to whom the file belongs. Logical, isn’t it? 😉
Yet another thing that doesn’t work the way as you would probably expect. At least it certainly didn’t for me. When I first discovered batching I thought — okay maybe hoped — that a batch request would count only as one towards the rate limit. It would have been heaven. Obviously, that is not the case. I know, I shouldn’t dream. Okay, my next instinct was that it is only there to alleviate the overhead of executing HTTP requests and it counts as it would normally. Even the documentation explicitly states that now. Doesn’t matter. It’s somewhere in the middle. It increases throughput, but not in a linear or consistent way. If you still had no need for a system described in the “Rate limiting” section you are rather lucky as you could utilize this to a certain degree. On the other hand, if you already have it and you are aiming at consistent “as-big-as-possible” throughput then it’s one more unknown factor in your equation. GL&HF.
It was rather easy. Just get a sufficiently big folder structure with enough files in it, which is shared with a lot of users/groups… and then start modifying things a lot. Every shared file has subscribers listening for changes. Anyway the more people it is shared with, the more of those exist. When you do a single API call, it might trigger a lot more events in the backend. If you do a LOT of API calls, well that is how you get weeks/months long queues on the Drive backend.
Imagine you have a resource which has a list property and you try to update/patch that property. I’m pretty sure you have the concept of how it should behave for empty or missing values. For example with an empty list provided what would you expect? I would expect the list to be cleared. Well that is not how drive.files.update/patch works for file.parents. Simply nothing happens. Rather confusing given that if you provide at least one parent value it works as anyone would expect. I know it has “addParents” and “removeParents” optional query parameters, but FFS why the inconsistent behavior, and why does it force the developers to get the current parent list before clearing it out? Not to mention there isn’t a single warning or documentation anywhere mentioning that it won’t work like that. It just silently skips it. The very same issue happens with drive.file.copy. I would expect to create it without a parent if I provide an empty list, but it makes no difference. You literally have to do another call to remove every parent the newly copied file has. Makes sense, doesn’t it? 😄
The following has been fixed for a while now, but they definitely worth mentioning. You know just to realize how much better it is now. 🙂
There are obvious response codes like ‘404 Not Found’. Some of them mean it shouldn’t be even retired. A part of them were to be retired immediately in the same request. Some of them meant to retry them later. Handling those were never an issue. Documentation regarding the matter was a lot more scarce that it is now. So, figuring out what to handle was a trial-and-error and logging what unhandled responses you get. It was not convenient, but eventually, you could cover the errors you usually get. And then, just because why not, they changed a few of those undocumented response codes without any warning or documentation at all. Just because why not. We were not happy. At all.
There was a time when any existing file was accepted as the parent. Even documents. It was a nice way to hide files from prying eyes, although I’m sure it wasn’t intended. 🙂 Also files’ ownership could have been transferred to groups and eventually completely lost. 🙂 Anyway both of these bugs have been fixed since.
Fortunately, this seems the be a thing of the past, but if you have ever received an error with the Document List API, I’m sure you will never forget it. Getting a whole HTML page inside the response. Pure epicness.
Even the much better/newer Google Drive API deserves an honorary mention. I mean who wouldn’t want to read error logs with Japanese messages 🙂
As you have seen I have listed a lot of issues, missing features, odd behaviors and trust me… there is more. 🙂 … and I still prefer working on apps running on GCP using/heavily relying on the Google Drive API over many other alternatives I could be doing. Obviously big part of that subjective view comes from GCP, so it’s not all about Drive. 🙂
And despite my long ranting most of these issues have been addressed somehow already. Some of them were fixed by Google, but for most of them, we managed to find some kind of a workaround. You know, sometimes we actually have to work for our salaries, and it’s not always about waiting for compilation/deployment/tool execution😉