Compare commits
9 Commits
bdcba003fe
...
write-read
Author | SHA1 | Date | |
---|---|---|---|
af45d1a9c8 | |||
1570c4559f | |||
5e56263469 | |||
4b46a735cf | |||
05dc8b41e1 | |||
02067139e6 | |||
259e19f2ac | |||
800712c63c | |||
721a38435b |
101
README.md
Normal file
101
README.md
Normal file
@@ -0,0 +1,101 @@
|
||||
|
||||
<!--Your Github repository must have the following contents:
|
||||
|
||||
A README.md file that communicates the libraries used, the motivation for the project, the files in the repository with a small description of each, a summary of the results of the analysis, and necessary acknowledgments.
|
||||
|
||||
Your code in a Jupyter notebook, with appropriate comments, analysis, and documentation.
|
||||
|
||||
You may also provide any other necessary documentation you find necessary.-->
|
||||
|
||||
# stacksurvey
|
||||
|
||||
**stacksurvey** is an exploration and analysis of data from StackOverflow's developer survey of 2024.
|
||||
|
||||
[https://survey.stackoverflow.co/2024/](https://survey.stackoverflow.co/2024/)
|
||||
|
||||
The motivation for project is satisfying a class assignment. Eventually, an interesting (enough) topic was discovered in the data set:
|
||||
|
||||
>What is the annual compensation (y) over years of experience (x) of deveopers who use a programming language from a specific country?
|
||||
|
||||
## Requirements
|
||||
|
||||
numpy pandas sklearn matplotlib seaborn
|
||||
|
||||
## Summary of Analysis
|
||||
|
||||
The models generated by the notebook become less reliable with years of experience greater than 10 or annual incomes greater than $200,000.
|
||||
|
||||
Each chart comes with two regression lines. Red is the default regression line that has not been tuned. The other is an attempt to better fit the data by either transforming or shifting x.
|
||||
|
||||
The transformation is typically
|
||||
|
||||
y = m * log(x) + b
|
||||
|
||||
where the base is a parameter.
|
||||
|
||||
Each model had different changes of base applied to the log function.
|
||||
|
||||
### C
|
||||
|
||||

|
||||
|
||||
+----------------------+
|
||||
red regression line for C
|
||||
coefficient = 1427.58
|
||||
intercept = 103659.82
|
||||
rmse = 26971.44
|
||||
r2 score = 0.06
|
||||
sample predictions:
|
||||
[[125073.46117519]
|
||||
[107942.54574181]
|
||||
[109370.12202793]]
|
||||
+----------------------+
|
||||
+----------------------+
|
||||
magenta regression line for C
|
||||
coefficient = 11973.47
|
||||
intercept = 54776.27
|
||||
rmse = 21198.61
|
||||
r2 score = 0.57
|
||||
sample predictions:
|
||||
[[132396.26294684]
|
||||
[119937.35465744]
|
||||
[ 64985.1549115 ]]
|
||||
+----------------------+
|
||||
|
||||
For C programmers, a linear model fits well but not great having an r2 score of 0.57. Junior level positions earn roughly $54,776. Their income progresses $11,973 with each year of experience.
|
||||
|
||||
### Python
|
||||
|
||||

|
||||
|
||||
+----------------------+
|
||||
red regression line for Python
|
||||
coefficient = 2573.62
|
||||
intercept = 123479.15
|
||||
rmse = 39759.45
|
||||
r2 score = 0.34
|
||||
sample predictions:
|
||||
[[126052.77118246]
|
||||
[174951.60602361]
|
||||
[187819.7204555 ]]
|
||||
+----------------------+
|
||||
+----------------------+
|
||||
cyan regression line for Python
|
||||
coefficient = 10378.53
|
||||
intercept = 82957.69
|
||||
rmse = 42374.26
|
||||
r2 score = 0.38
|
||||
sample predictions:
|
||||
[[139882.01866593]
|
||||
[117229.55243376]
|
||||
[137277.30441955]]
|
||||
+----------------------+
|
||||
|
||||
For data scientists, analysts, or engineers, a linear model is a moderate fit at best as the r2 score is around 0.30. There appears to be divergence at the 10 year mark in their careers. This may be the result of their field (advertising, finance, bio/medical, and so on).
|
||||
|
||||
Entry or junior level professionals generally have an income of $82,957 or $123,479. Their annual income increases by $10,378 or $2573 each year.
|
||||
|
||||
## Acknowledgements
|
||||
|
||||
* "Udacity AI" (ChatGPT), the idea to transform x values to appropriate a linear regression into a logarithmic regression.
|
||||
|
BIN
images/programmers-C-United-States-of-America.png
Normal file
BIN
images/programmers-C-United-States-of-America.png
Normal file
Binary file not shown.
After Width: | Height: | Size: 81 KiB |
BIN
images/programmers-Python-United-States-of-America.png
Normal file
BIN
images/programmers-Python-United-States-of-America.png
Normal file
Binary file not shown.
After Width: | Height: | Size: 91 KiB |
@@ -149,22 +149,12 @@
|
||||
" self.devs = devs.drop(indices)\n",
|
||||
" del devs, new_column\n",
|
||||
" \n",
|
||||
" def visualize(self, n_lowest=0, \n",
|
||||
" hue=\"Country\", palette=sb.color_palette() ): \n",
|
||||
" def visualize(self, hue=\"Country\", \n",
|
||||
" palette=sb.color_palette() ): \n",
|
||||
" self.canvas = plt.figure()\n",
|
||||
" key_x = self.key_x\n",
|
||||
" key_y = self.key_y\n",
|
||||
"\n",
|
||||
" if n_lowest > 0:\n",
|
||||
" # chatgpt draws my line\n",
|
||||
" # Calculate the lowest nth point (for example, the 5th lowest value)\n",
|
||||
" # iloc[-1] gets the last element from the n smallest\n",
|
||||
" lowest_nth = self.devs[key_y].nsmallest(n_lowest).iloc[-1] \n",
|
||||
" # Draw a horizontal line at the lowest nth point\n",
|
||||
" # label=f'Lowest {n_poorest}th Point: {lowest_nth_value:.2f}'\n",
|
||||
" plt.axhline(y=lowest_nth, color='purple', linestyle='--', \n",
|
||||
" label=\"y=%0.2f\" % lowest_nth, zorder=-1 )\n",
|
||||
"\n",
|
||||
" sb.scatterplot(data=self.devs, x=key_x, y=key_y, hue=hue, palette=palette)\n",
|
||||
" plt.legend(loc='lower center', bbox_to_anchor=(1.5,0)) \n",
|
||||
" title = \"Annual Salary of %s Developers Over Years of Experience\" % self.language\\\n",
|
||||
@@ -200,10 +190,12 @@
|
||||
" print(\"sample predictions:\")\n",
|
||||
" print(y_pred[3:6])\n",
|
||||
" print(\"+----------------------+\")\n",
|
||||
" \n",
|
||||
" plt.figure(self.canvas)\n",
|
||||
" b = model.intercept_[0]\n",
|
||||
"\n",
|
||||
" plt.figure(self.canvas)\n",
|
||||
" plt.plot(X_test, y_pred, color=line_color, label='Regression Line')\n",
|
||||
" plt.axhline(y=b, color=\"purple\", linestyle='--', \n",
|
||||
" label=\"b=%0.2f\" % b, zorder=-1 )\n",
|
||||
" plt.legend(loc='lower center', bbox_to_anchor=(1.5,0)) \n",
|
||||
" del y_pred, model\n",
|
||||
"\n",
|
||||
@@ -280,7 +272,7 @@
|
||||
" # \"Product manager\"\n",
|
||||
"]\n",
|
||||
"c = Foo(so_df, \"C\", jobs=cjobs, n_rich_outliers=30, n_poor_outliers=2)\n",
|
||||
"c.visualize(n_lowest=7, hue=\"DevType\", palette=[\"#57e6da\",\"#d9e352\",\"#cc622d\"] ) \n",
|
||||
"c.visualize(hue=\"DevType\", palette=[\"#57e6da\",\"#d9e352\",\"#cc622d\"] ) \n",
|
||||
"c.run_regression()\n",
|
||||
"c.run_regression(x_transform=log_base_a, change_base=1.3, \n",
|
||||
" x_shift=2, y_shift=-5000, line_color=\"magenta\", random=555)\n",
|
||||
|
Reference in New Issue
Block a user