A list of puns related to "Predictive Modeling"
I've been with a mid-sized P&C company for three years. Most of my work was data support + building/maintaining SAS raters for indications / on-leveling / backing into factors / competitive analysis. I'm one exam away from ACAS, but I feel I haven't done much traditional reserving or ratemaking. The chief wants to split apart with me and perform predictive modeling (MAS-I and MAS-II material), and I can keep my exam support. I like the material from those exams and find modeling interesting. I worry that not having more traditional ratemaking / reserving will hurt future job opportunities. Any thoughts /r/actuary?
Hi everyone!
In a recent post, I detailed some work I had been doing on predicting whether reviewers would add upcoming games to their collection. Iβve since updated that project so that I can take anyone with a collection on BGG (Iβve gotten solid results for folks with as few as 30 games) on BGG and spit out an analysis like so for your collection..
This post is about a different but related analysis: predicting the geek (and average) rating for upcoming games. Iβm still tinkering with the models, but Iβd like to share some of the early results with the community to gather feedback and to also facilitate some discussion about additions to the model.
Motivation:
Can we predict which upcoming games will be the most highly rated on BGG?
Letβs wind the clock back to the end of 2018. Could we have predicted which 2019 and 2020 (and so on) games would go on to become the top rated games on the geek list?
Upcoming games enter the BGG database well before they are published, so we already have a lot of information to work with in estimating newly released games - we know categories, designers, publishers, mechanics, artists, etc. Every game starts out at 5.5 on the geek rating, and starts to slowly shift if a game accumulates enough ratings - games that achieve high geek ratings need to have 1) lots of user ratings and 2) a high average rating. We wonβt know a gameβs geek rating for a while, but we can train a model to estimate it for us. We have plenty of historical data on games and their geek rating, we can use a model to learn the relationship between a gameβs features and its geek rating, which we can then use to predict new games.
Examples:
To illustrate how this works, letβs look at a past example. On Mars first showed up in the historical BGG data on January 24, 2019. At that point, we knew its publisher, its mechanics, its designer, its artist - we all probably had a decent guess that it was going to be a popular game just from the latter two alone. And indeed [On Mars has steadily climbed the geek list s
... keep reading on reddit β‘My company is trying to better estimate and time the actual cash collected from our commercial customer on a DAILY basis for the coming quarter. We have about 6 months of data to work with. But it seems there can be some serious variability and its quite hard to predict.
I was thinking of implementing an ARIMA model in python but was curious if there are any other predictive modeling techniques anyone has used for similar problems?
I have tried naive and simple moving average but it has proven inaccurate.
https://preview.redd.it/3e1o45f9dyq71.jpg?width=2000&format=pjpg&auto=webp&s=f587b3a5a8cf40e269197d0a802f01a0b42c6305
TomTom Maps and API services produce massive volumes of data. Data scientists can access this data to gain insight and make predictions supporting business decisions for better performance and customer satisfaction.
We data scientists and developers can find various historical and real-time data to help with our projects, such as traffic stats, real-time traffic, location history, notifications, maps and routing, and road analytics. TomTomβs extensive documentation and developer community support us as we play with this rich, easily accessible data. For example, we can use TomTomβs data for predictive modeling.
In this article, weβll do a hands-on exploration of how to use TomTom API data for predictive modeling. Weβll use accident data to learn when and where accidents are most likely to take place. Youβll see just how easy it is to use TomTomβs mapping data for your data science projects. Weβll pull traffic data from a map application, which connects to TomTom APIs, then extract a dataset to build a model for predictions. For our supervised learning task, weβll be using RandomForestRegressor.
After training the model, weβll evaluate and refine it until it is accurate, then deploy it to work with the TomTom Maps API in real-time. Finally, weβll use Python and some favorite data science tools (Jupyter Notebook, NumPy, Pandas, and scikit-learn) to explore our data and make predictions.
https://preview.redd.it/d0p0eu0edyq71.jpg?width=468&format=pjpg&auto=webp&s=ded1ff8601917e3d9e74d463eed6f35e703d94b1
To create our model, we first pull data from TomTom APIs connected to a map application. Then, we follow the framework in the image below to prepare the dataset, build, train, and evaluate the model. If necessary, we refine our model. Finally, we deploy our model to deliver t
... keep reading on reddit β‘I have offers in both EL pricing and EL predictive modeling work. I would greatly appreciate any advice/input in deciding which to pursue. Some of my (possibly dumb/obvious) questions/concerns:
Thanks in advance!
As the title says, I am trying to gauge interest for a project that I have had floating around in my head for a while. The reason I am interested is that I want to provide enterprise-level tools for small ops and everyday growers at a minimal cost level.
I want the information to be open-sourced and made available. I, however, would need an enormous amount of data from small-scale growers to generate the training models. I have been looking at using growdaries as a possible place to gather some of this data but need to make sure it is within terms of service.
That's why I am posting this here. If there is enough interest, I can set up a place to drop and centralize data. I know data is going to be the backbone of future growth and I understand how valuable that can be, so asking to make it open-source is a huge leap. But the more accessible the tool, the more people that can utilize it, keeping competition rich and the art of growing away from large corps.
Post below with questions, blasts, or opinions. Hope every is staying safe and healthy
Looking for someone who has some experience and wouldn't mind having a private conversation about what they do. Trying to decide if I'd prefer that flavor of work over traditional actuarial work.
My Perspective:
Hi all,
I'm an aspiring sports AI/ML data analyst and am writing a blog to showcase and practice analytical modeling work. My most recent post delves into NBA Betting Models for spread, total points, and FanDuel DFS lineups. Check it out:
I have been practicing with Kaggle competitions and am now trying to answer a real problem for a friend and am immediately stuck. The data is in two tables:
- One table has 1 row per user per day. It has the advertisement the user watched to install the game , the revenue from the user per day, and the age of the user on the day
- The other table is a list of each advertisement and how much it cost to run that ad
So as an easy example, let's say "Ad 1" cost him $3 to run and "Ad 2" cost him $5 to run. And the user table is:
username | day | ad_watched | daily_revenue | user_age |
---|---|---|---|---|
alice | monday | Ad 1 | $1 | 1 |
bob | monday | Ad 1 | $1 | 1 |
alice | tuesday | Ad 1 | $1 | 2 |
bob | tuesday | Ad 1 | $1 | 2 |
carl | tuesday | Ad 2 | $5 | 1 |
alice | wednesday | Ad 1 | $1 | 3 |
bob | wednesday | Ad 2 | $1 | 3 |
carl | wednesday | Ad 2 | $2.50 | 2 |
If I was modelling this by hand on Thursday, I would see that "Ad 1" cost $3 and brought in 2 users who seem to pay $1 per day consistently: their projected revenue would be $2 for Thursday. And "Ad 2" cost $5 and brought in one user and if the trend continues I expect him to bring in $1.25 on Thursday. So Thursday's expected revenue is $3.25 (of course this is a massive exaggeration but I'm just trying to get across what I'm trying to do).
Basically, I want to train a model to (1) predict the expected daily revenue per day for the next few weeks and (2) predict the revenue for each Ad type. Ideally by blending with the other Ad's data (for example, I haven't ran Ad 2 for many months but the other Ads have been performing better due to game improvements. I would like the model to follow the trend of Ad 2 in the past but lifted by the amount the other Ads have recently lifted). I am confused because my actual model should predict a simple "per day, per Ad" expected revenue, but that looks nothing like my tables and I'm not sure how to create a new table to feed it without losing a lot of data in the process.
I am having trouble visualizing how to train the model or what I am really doing. I could really use a nudge in the right direction. Thanks!
I want to get a book that details the entire modeling process to patch up some gaps I may have. I keep coming back to Applied Predictive Modeling but it was published in 2013. (From what I can tell, the 2018 edition isn't updated.) R has changed a lot since then and I'd like to read something that assumes a tidyverse-oriented workflow.
Anyone have any ideas, or is that book just the gold standard? R4DS is great but is far more about R than about the DS and I'm really wanting something that focuses on the latter. Thanks y'all!
Hello everyone!
This year I found a very interesting piece of software that will enable a designer to predict a userβs time on task at design-time.
Basically, you mockup a design (just as you would in Balsamiq or Sketch), then demonstrate a task (tell CogTool what the user would do). Then the software predicts how long it would take a human to perform that task. (using GOMS / ACT-R) The best part is the tool is completely visual.
Download: www.cogtool.org
This seems like very powerful tool!
Iβve taken this open source software project under my wing. If you find it useful, please share your experience. (If you find bugs, please let me know in the comments or via our GitHub)
I know how to use R, but need to learn predictive modeling/analysis relatively quickly. Thanks!
Please note that this site uses cookies to personalise content and adverts, to provide social media features, and to analyse web traffic. Click here for more information.