Frequently asked questions
The number of Tweets I receive with the data endpoint doesn't match the number of Tweets identified by the counts endpoint. Why is this the case?
There is a known difference between results provided by the counts endpoint and the data endpoint. You may see a discrepancy in your results because the counts endpoint is pre-compliance (meaning that it does not account for things like deleted Tweets, scrub geo, etc.) whereas the data endpoint is compliant at the time of delivery and accounts for all compliance events. For further reference, please go to this document on our support site.
I didn't receive a Tweet that should match my query. Why?
There are a few different reasons why this might have happened, including
- the Tweet you expected to see is from a protected account
- because the data endpoint accounts for all compliance events (meaning that deleted Tweets, scrubbed geos, etc. will not be included in the response).
My query matched a Tweet but includes a keyword that I negated. Why is this happening?
This is likely due to a wrong usage of our premium rules & filtering. Please review our documentation here and ensure you understand the restrictions around building rules.
Are there any libraries that I can use to get started using the Search Tweet APIs?
Yes, there are, including:
- Tweepy - good for using the standard search/tweets product (Python)
- Twitter API - good for using both the standard and premium Search Tweet APIs (Python)
- Search Tweets Python and Search Tweets Ruby - two good tools that can be used for both premium and enterprise Search Tweet APIs
All of the libraries that we directly support can be found on our TwitterDev GitHub page: https://github.com/twitterdev.
There are other third-party libraries that may also be helpful; however, please note that some of these may not work with our premium and enterprise products.
Will I ever receive less volume of Tweets than the value I set as the
maxResults in my request to the data endpoint?
The data endpoint paginates at the specified
maxResults as long as there are enough matches. The last page of results may have less data than maxResults.
The counts endpoint paginates at 31 days. For example, setting a day bucket, will return the count of results per day for 31 days per page. Setting an hour bucket, will return the count of results per hour for 744 (31 days x 24 hours) hours per page.
In what order are the matching Tweets returned?
Tweets are returned in reverse chronological order. For example, the first page of results will show the most recent Tweets that match the query, pagination will continue until the results posted dates get to the
fromDate initially requested.
I'm interested in learning more about the pricing of the enterprise Search Tweet API and in applying for this offering. How can I do this?
Our enterprise solutions are customized with predictable pricing to meet the needs of your business. Please apply here for more information.
How do I build a rule set that matches my use case?
- Please refer to our enterprise Search Tweet APIs documentation here
- Useful information on rules and filering can be found here
- Useful information for using the data endpoint can be found here
- Useful information for using the counts endpoint can be found here
- A list of available operators can be found here
I have exceeded my request caps/limits for the month, but I need to access more data - what can I do?
Please get in touch with your Account Manager at Twitter who will be able to help you with this.
Code 404 - Not Found
- Please ensure you are using the right parameters for each endpoint (e.g. the
bucketsfield can only be used with the counts endpoint, not the data endpoint)
- Please double check the
:labelfields are correct. You can find your
:labelfield in the GNIP Console (enterprise customers only).