|
8 | 8 | import random
|
9 | 9 | from scipy.stats import norm
|
10 | 10 |
|
11 |
| -# data = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/df2020.csv') |
12 |
| -# df2018 = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/df2018.csv') |
13 |
| -# full_data2018 = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/survey_results_sample_2018.csv') |
14 |
| -# full_data2019=pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/survey_results_sample_2019.csv') |
15 |
| -# full_df2020 = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/survey_results_sample_2020.csv') |
16 |
| -# df2019 = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/df2019.csv') |
17 |
| -# df2020 = data[(data['SalaryUSD'] < 200000)] |
| 11 | +data = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/df2020.csv') |
| 12 | +df2018 = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/df2018.csv') |
| 13 | +full_data2018 = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/survey_results_sample_2018.csv') |
| 14 | +full_data2019=pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/survey_results_sample_2019.csv') |
| 15 | +full_df2020 = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/survey_results_sample_2020.csv') |
| 16 | +df2019 = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/df2019.csv') |
| 17 | +df2020 = data[(data['SalaryUSD'] < 200000)] |
18 | 18 |
|
19 |
| -# # features for job satisfaction |
20 |
| -# results = pd.read_csv("https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/results.csv") |
| 19 | +# features for job satisfaction |
| 20 | +results = pd.read_csv("https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/results.csv") |
21 | 21 |
|
22 | 22 |
|
23 | 23 | # Load the data from local CSV files
|
24 |
| -data = pd.read_csv('df2020.csv') |
25 |
| -df2018 = pd.read_csv('df2018.csv') |
26 |
| -full_data2018 = pd.read_csv('survey_results_sample_2018.csv') |
27 |
| -full_data2019 = pd.read_csv('survey_results_sample_2019.csv') |
28 |
| -full_df2020 = pd.read_csv('survey_results_sample_2020.csv') |
29 |
| -df2019 = pd.read_csv('df2019.csv') |
30 |
| - |
31 |
| -# Filter the 2020 data |
32 |
| -df2020 = data[(data['SalaryUSD'] < 200000)] |
| 24 | +# data = pd.read_csv('df2020.csv') |
| 25 | +# df2018 = pd.read_csv('df2018.csv') |
| 26 | +# full_data2018 = pd.read_csv('survey_results_sample_2018.csv') |
| 27 | +# full_data2019 = pd.read_csv('survey_results_sample_2019.csv') |
| 28 | +# full_df2020 = pd.read_csv('survey_results_sample_2020.csv') |
| 29 | +# df2019 = pd.read_csv('df2019.csv') |
| 30 | + |
| 31 | +# # Filter the 2020 data |
| 32 | +# df2020 = data[(data['SalaryUSD'] < 200000)] |
33 | 33 |
|
34 |
| -# Load results for job satisfaction from the local file |
35 |
| -results = pd.read_csv("results.csv") |
| 34 | +# # Load results for job satisfaction from the local file |
| 35 | +# results = pd.read_csv("results.csv") |
36 | 36 |
|
37 | 37 |
|
38 | 38 | #######################################
|
|
0 commit comments