bearer_token()
docs.list_statuses()
endpointnetwork_data()
and network_graph()
lookup_coords()
now requires a Google Maps API key. It will be stored for easy future use once supplied.bearer_token()
option for access to more generous rate limits.create_token()
when using browse-based authentication method.post_list()
, which now allows users to create and populate lists as well as delete lists on behalf of one’s own Twitter account.lists_memberships()
and now scrolls through multiple pages of results to automate collection of larger numbers of lists.create_token()
which allows for creation of token non-interactive sessions via accepting inputs for consumer key, consumer secret (always required), oauth key, and oauth secret (optional, if supplied then non-browser sign method is used).ts_*()
functions now offer a tz
(timezone) argument, allowing users to more easily print and plot in non-UTC time.destroy_id
argument in post_tweet()
join_rtweet()
, which omitted users who didn’t have available tweets.all_suggested_users()
, which automates the collection of Twitter’s suggested users data.save_as_csv()
, including addition of new prep_as_csv()
as convenience function for flattening Twitter data frames.save_as_csv()
with improved flattening and ID-preserving saving methods. The function now saves a single [joined] data set as well.get_favorites()
and in several lists_*()
functions.stream_tweets()
stream_tweets2()
function for more robust streaming method. Streams JSON files to directory and reconnects following premature disruptions.get_timeline()
, get_favorites()
, get_friends()
, and get_followers()
now accept vectors of length > 1.users_data()
stream_data()
, designed to parse files that cannot wholely fit into memory. stream_data()
can now work in parallel as well.post_status()
function has been fixed and can now be used to upload media.mentions_screen_name
may consist of 4 screen names).flatten()
function. Exporting functions such as save_as_csv
will apply flatten by default.coords_coords
, geo_coords
, and bbox_coords
bounding box. The first two come in pairs of coords (a list column) and bbox_coords comes with 8 values (longX4 latX4). This should allow users to maximize returns on geo-location data.plain_tweets()
added for textual analysis.ts_plot()
with improved time-aggregating method. Now a wrapper around ts_data()
, deprecating ts_filter
.mutate_coords()
and retryonratelimit
.ts_plot()
, ts_filter()
and more!post_
. This was done to clearly distinguish write functions from retrieval functions.ts_plot()
function is now more robust with more adaptive characteristics for variations in the number of filters, the method of distinguishing lines, the position of the legend, and the aesthetics of the themes.ts_filter()
function which allows users to convert Twitter data into a time series-like data frame. Users may also provide filtering rules with which ts_filter()
will subset the data as it converts it to multiple time series, which it then outputs as a long-form (tidy) data frame.search_tweets
now includes retryonratelimit
argument to allow for searches requesting more than 18,000 tweets. This automates what was previously possible through use of max_id
.stream_tweets
.parse.piper
functionts_plot
to enable different filtered time series and an aesthetic overhaul of the plot function as well.as_double
argument to provide flexibility in handling id variables (as_double provides performance boost but can create problems when printing and saving, depending on format). By default functions will return IDs as character vectors.clean_tweets
argument provided to allow user more control over encoding and handling of non-ascii characters.search_users
and implemented several improvements to stream_tweets
and plot_ts
.twitter_tokens
, twitter_token
, tokens
, or token
, rtweet will find it.search_tweets
and stream_tweeets
include_retweets
arg added to search_tweets()
function.user_id
class changed to double when parsed. double is significantly faster and consumes less space. it’s also capable of handling the length of id scalars, so the only downside is truncated printing.stream_tweets()
stream_tweets()
function. By default, the streaming query argument, q
, is now set to an empty string, q = ""
, which returns a random sample of all Tweets (pretty cool, right?).post_tweet()
function. Users can now post tweets from their R console.get_favorites()
functionAdded lookup_statuses()
function, which is the counterpart to lookup_users()
. Supply a vector of status IDs and return tweet data for each status. lookup_statuses()
is particularly powerful when combined with other methods designed to collect older Tweets. Early experiments with doing this all through R have turned out surprisingly well, but packaging it in a way that makes it easy to do on other machines is unlikely to happen in the short term.
Removed dplyr dependencies. Everyone should install and use dplyr
, but for sake of parsimony, it’s been removed from rtweet.
Continued development of S4 classes and methods. Given removal of dplyr dependencies, I’ve started to integrate print/show methods that will limit the number of rows (and width of columns) when printed. Given the amount of data returned in a relatively short period of time, printing entire data frames quickly becomes headache-inducing.
Added new trends functions. Find what trending locations are available with trends_available()
and/or search for trends worldwide or by geographical location using get_trends()
.
Stability improvements including integration with Travis CI and code analysis via codecov. Token encryption method also means API testing conducted on multiple machines and systems.
search_users()
function! Search for users by keyword, name, or interest and return data on the first 1000 hits.Output for search_tweets()
, stream_tweets()
, and get_timeline()
now consists of tweets data and contains users data attribute.
Output for lookup_users()
now consists of users data and contains tweets data attribute.
To access users data from a tweets object or vice-versa, use users_data()
and tweets_data()
functions on objects output by major rtweet retrieval functions.
Updated testthat tests
Output for get_friends()
and get_followers()
is now a tibble of “ids”. To retrieve next cursor value, use new next_cursor()
function.
Major stability improvements via testthat tests for every major function.
Since previous CRAN release, numerous new features and improvements to functions returning tweets, user data, and ids.
Search function now optimized to return more tweets per search.
Numerous improvements to stability, error checks, and namespace management.
Improvements to get_friends
and get_followers
. Returns list with value (next_cursor
) used for next page of results. When this value is 0, all results have been returned.
Functions get_friends
and get_followers
now return the list of user ids as a tibble data table, which makes the print out much cleaner.
Improved scrolling methods such that search_tweets
and get_timeline
should return a lot more now
Added parser
function to return status (tweets) AND user (users) data frames when available. As a result, the parsed output for some functions now comes as a list containing two data frames.
Added get_timeline
function that returns tweets from selected user
Added vignettes covering tokens and search tweets
Fixed issue with count
argument in search and user functions
Fixed parsing issue for return objects with omitted variables
Added clean_tweets
convenience function for text analysis
More examples included in documentation.
Added recode_error
argument to get_friends
function. This is especially useful for tracking networks over time.
Further integrated ROAuth
methods/objects to increase compatibility with twitteR
authorization procedures.
Improved token checking procedures.
Added NEWS.md
file
Added key features
and more descriptions to README.md
.
There are now two stable parse (convert json obj to data frame) types. For user objects (e.g., output of lookup_users
), there is parse_user
. For tweet objects (e.g., output of search_tweets
or stream_tweets
), there is parse_tweets
.
New parse functions are now exported, so they should available for use with compatible Twitter packages or user-defined API request operations.
More parsing improvements
Added format_date
function
Various stability improvements