-
Notifications
You must be signed in to change notification settings - Fork 915
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
dnsdist: cache miss when all tcp only backends are down #15337
Comments
I'll think about it a bit more, but right now I think what we are currently doing is the only safe approach with the information we have: the query has been received over UDP, we have no information about whether the server is TCP-only, so we do a UDP-based lookup.
|
I can see a manual flag go wrong indeed, but does an automatically set flag: "all servers in this pool are TCP only" have the same drawback? This flag should be re-computed on each pool modification. |
A few points in my mind for your considerations:
I guess the main focus is that, if an upstream is udp, is it safe to find a cache result inserted from a tcp backend? I would think it is safe If we are not blindly sending back cache result as a response. |
Yes, otherwise we need to go the backend for every UDP query. Even if we assume that the backend is eventually going to give us a response over TCP (and it might never happen), if it's larger than the EDNS UDP payload size (or 512 if there is no EDNS in the query) we will have to serve the truncated answer. We used to not cache empty queries and it caused a lot of problems.
Once we get the TCP answer, which might never happen. Also note that we might get a UDP response from the backend that is not truncated but different than the TCP one, because all required records fit but some optional ones did not, so in this case we need to keep the UDP answer as otherwise we would have to tell UDP clients to retry over TCP because the response does not fit, degrading performance. |
I can see that things get too complicated, with including additional RR parts into cache key calculation. I ran into another issue that when I run a query with EDNS payload 800, while actual result is 1000, so the downstream server returns result with TC. Then the client auto switched to tcp and get the correct result. However, all next the same queries will always get cache miss and can only be relayed to the downstream server first, only after get a TC from the server, then switch to tcp, and can hit the cache and get a result back. Is this reasonable? In this case, if it is too complicated and error prone to provide a single solution for all possible use case combinations, then we may consider providing configurable options so user can choose a cache key strategy for its own use case. For example, the very basic strategy is to only use dnsheader/qname/class/type, then other parts can be added based on user's configuration, so additional RRs or transport type can be configurable. |
And I did a quick test that, for the same udp query, when I do followings:
Each of above query will create a new cache entry. I don't think this is a quite reasonable behavior. |
What dnsdist has is a packet cache, which has some caveats like this one. On the other hand it's very fast, and doesn't aim to replace more complicated caches that the backend pretty much always has, especially since the backend has much more information about the response. So we need to weight the benefits of making the cache more complex, because if in a non-negligible percentage of the cases it slows down the processing compared to no packet cache at all, it's a net loss. So, correctness issues are not acceptable and we will fix them. Misses that could be turned into hits, these are possible improvements that will be implemented if and only if they don't significantly increase the complexity. |
This is due to that, the original udp query only has client cookie, while the fall back tcp query cotains both client and server cookie. Though the cookie content is skipped for cache key calculation, the RR data length and cookie option length both are used. So the fall back tcp query always has a different key than the original udp query. |
Ok, agree that my purposed designed test cases with different payload size is likely not realistic. In real world, one client is unlikely to change payload size dynamically. Multiple clients may be, while usually different client programs will be interested in different QNAME as well. |
Short description
Environment
Compiled by myself according to steps in repository builder directory
Steps to reproduce
Expected behaviour
Expect cache hit happens as a result of step 8
Actual behaviour
Currently cache miss is happening
Other information
After digging it seems this is related with cache key calculation strategy in dnsdist. More details can be found in #15335
The text was updated successfully, but these errors were encountered: