[{"title":"( 182 个子文件 33.77MB ) scikit-learn-Cookbook-Second-Edition:Packt发行的scikit-learn Cookbook第二版","children":[{"title":"dtree.clf <span style='color:#111;'> 2.54KB </span>","children":null,"spread":false},{"title":"Writing+A+Stacking+Agreggator+with+Scikit-Learn+Extra(2).html <span style='color:#111;'> 284.74KB </span>","children":null,"spread":false},{"title":"Tuning a Decision Tree.ipynb <span style='color:#111;'> 892.89KB </span>","children":null,"spread":false},{"title":"Multiclass SVC Classifier.ipynb <span style='color:#111;'> 663.88KB </span>","children":null,"spread":false},{"title":"Quantizing an image with KMeans clustering-checkpoint.ipynb <span style='color:#111;'> 277.20KB </span>","children":null,"spread":false},{"title":"Quantizing an image with KMeans clustering.ipynb <span style='color:#111;'> 277.20KB </span>","children":null,"spread":false},{"title":"Optimizing a Support Vector Machine-checkpoint.ipynb <span style='color:#111;'> 204.05KB </span>","children":null,"spread":false},{"title":"Optimizing a Support Vector Machine.ipynb <span style='color:#111;'> 204.05KB </span>","children":null,"spread":false},{"title":"Classifying data with a Linear Support Vector Machines-checkpoint.ipynb <span style='color:#111;'> 191.09KB </span>","children":null,"spread":false},{"title":"Classifying data with a Linear Support Vector Machines.ipynb <span style='color:#111;'> 191.09KB </span>","children":null,"spread":false},{"title":"Reducing dimensionality with PCA-checkpoint.ipynb <span style='color:#111;'> 180.64KB </span>","children":null,"spread":false},{"title":"Reducing dimensionality with PCA.ipynb <span style='color:#111;'> 180.64KB </span>","children":null,"spread":false},{"title":"Dimensionality Reduction with Manifolds tSNE-checkpoint.ipynb <span style='color:#111;'> 148.48KB </span>","children":null,"spread":false},{"title":"Dimensionality Reduction with Manifolds tSNE.ipynb <span style='color:#111;'> 148.48KB </span>","children":null,"spread":false},{"title":"Using Gaussian processes for regression-checkpoint.ipynb <span style='color:#111;'> 128.50KB </span>","children":null,"spread":false},{"title":"Using Gaussian processes for regression.ipynb <span style='color:#111;'> 128.50KB </span>","children":null,"spread":false},{"title":"Visualize a Decision Tree with pydot-checkpoint.ipynb <span style='color:#111;'> 124.64KB </span>","children":null,"spread":false},{"title":"Visualize a Decision Tree with pydot.ipynb <span style='color:#111;'> 124.64KB </span>","children":null,"spread":false},{"title":"Using k-NN for regression-checkpoint.ipynb <span style='color:#111;'> 113.36KB </span>","children":null,"spread":false},{"title":"Using k-NN for regression.ipynb <span style='color:#111;'> 113.36KB </span>","children":null,"spread":false},{"title":"Plotting with Numpy and Matplotlib-checkpoint.ipynb <span style='color:#111;'> 108.00KB </span>","children":null,"spread":false},{"title":"Plotting with Numpy and Matplotlib.ipynb <span style='color:#111;'> 108.00KB </span>","children":null,"spread":false},{"title":"Using KMeans to cluster data-checkpoint.ipynb <span style='color:#111;'> 104.16KB </span>","children":null,"spread":false},{"title":"Using KMeans to cluster data.ipynb <span style='color:#111;'> 104.16KB </span>","children":null,"spread":false},{"title":"Pipelines Testing Methods to Reduce Dimensionality-checkpoint.ipynb <span style='color:#111;'> 95.27KB </span>","children":null,"spread":false},{"title":"Pipelines Testing Methods to Reduce Dimensionality.ipynb <span style='color:#111;'> 95.27KB </span>","children":null,"spread":false},{"title":"Assessing cluster correctness-checkpoint.ipynb <span style='color:#111;'> 93.17KB </span>","children":null,"spread":false},{"title":"Assessing cluster correctness.ipynb <span style='color:#111;'> 93.17KB </span>","children":null,"spread":false},{"title":"Decomposition to classify with DictionaryLearning-checkpoint.ipynb <span style='color:#111;'> 92.37KB </span>","children":null,"spread":false},{"title":"Decomposition to classify with DictionaryLearning.ipynb <span style='color:#111;'> 92.37KB </span>","children":null,"spread":false},{"title":"Viewing the Pima Indians Diabetes Dataset with Pandas-checkpoint.ipynb <span style='color:#111;'> 76.10KB </span>","children":null,"spread":false},{"title":"Viewing the Pima Indians Diabetes Dataset with Pandas.ipynb <span style='color:#111;'> 76.10KB </span>","children":null,"spread":false},{"title":"Taking a more fundamental approach to regularization with LARS-checkpoint.ipynb <span style='color:#111;'> 75.81KB </span>","children":null,"spread":false},{"title":"Taking a more fundamental approach to regularization with LARS.ipynb <span style='color:#111;'> 75.81KB </span>","children":null,"spread":false},{"title":"Probabilistic clustering with Gaussian Mixture Models.ipynb <span style='color:#111;'> 69.51KB </span>","children":null,"spread":false},{"title":"Doing basic classifications with Decision Trees.ipynb <span style='color:#111;'> 61.55KB </span>","children":null,"spread":false},{"title":"Using KMeans for outlier detection-checkpoint.ipynb <span style='color:#111;'> 59.90KB </span>","children":null,"spread":false},{"title":"Using KMeans for outlier detection.ipynb <span style='color:#111;'> 59.90KB </span>","children":null,"spread":false},{"title":"Using LDA for classification.ipynb <span style='color:#111;'> 58.32KB </span>","children":null,"spread":false},{"title":"A linear model in the presence of outliers.ipynb <span style='color:#111;'> 55.17KB </span>","children":null,"spread":false},{"title":"A linear model in the presence of outliers-checkpoint.ipynb <span style='color:#111;'> 55.17KB </span>","children":null,"spread":false},{"title":"Varying the Classification Threshold in Logistic Regression.ipynb <span style='color:#111;'> 45.89KB </span>","children":null,"spread":false},{"title":"Optimizing the ridge regression parameter.ipynb <span style='color:#111;'> 43.43KB </span>","children":null,"spread":false},{"title":"Using ridge regression to overcome linear regression's shortfalls-checkpoint.ipynb <span style='color:#111;'> 41.17KB </span>","children":null,"spread":false},{"title":"Using ridge regression to overcome linear regression's shortfalls.ipynb <span style='color:#111;'> 41.17KB </span>","children":null,"spread":false},{"title":"Tuning a Random Forest-checkpoint.ipynb <span style='color:#111;'> 40.76KB </span>","children":null,"spread":false},{"title":"Tuning a Random Forest.ipynb <span style='color:#111;'> 40.76KB </span>","children":null,"spread":false},{"title":"Evaluating the linear regression model.ipynb <span style='color:#111;'> 35.75KB </span>","children":null,"spread":false},{"title":"Tuning Gradient Boosting Trees.ipynb <span style='color:#111;'> 34.94KB </span>","children":null,"spread":false},{"title":"Decision Trees for Regression-checkpoint.ipynb <span style='color:#111;'> 34.55KB </span>","children":null,"spread":false},{"title":"Decision Trees for Regression.ipynb <span style='color:#111;'> 34.55KB </span>","children":null,"spread":false},{"title":"Cross-validation with ShuffleSplit.ipynb <span style='color:#111;'> 31.90KB </span>","children":null,"spread":false},{"title":"Tuning a Decision Tree-checkpoint.ipynb <span style='color:#111;'> 31.59KB </span>","children":null,"spread":false},{"title":"Writing A Stacking Agreggator with Scikit-Learn-checkpoint.ipynb <span style='color:#111;'> 27.15KB </span>","children":null,"spread":false},{"title":"Writing A Stacking Agreggator with Scikit-Learn.ipynb <span style='color:#111;'> 27.15KB </span>","children":null,"spread":false},{"title":"Finding the closest object in the feature space-checkpoint.ipynb <span style='color:#111;'> 25.28KB </span>","children":null,"spread":false},{"title":"Finding the closest object in the feature space.ipynb <span style='color:#111;'> 25.28KB </span>","children":null,"spread":false},{"title":"Scaling data to the standard normal.ipynb <span style='color:#111;'> 24.74KB </span>","children":null,"spread":false},{"title":"Random Forest Regression-checkpoint.ipynb <span style='color:#111;'> 23.25KB </span>","children":null,"spread":false},{"title":"Random Forest Regression.ipynb <span style='color:#111;'> 23.25KB </span>","children":null,"spread":false},{"title":"ROC Analysis-checkpoint.ipynb <span style='color:#111;'> 22.49KB </span>","children":null,"spread":false},{"title":"Creating sample data for toy analysis.ipynb <span style='color:#111;'> 21.78KB </span>","children":null,"spread":false},{"title":"Viewing the Iris Dataset with Pandas-checkpoint.ipynb <span style='color:#111;'> 21.59KB </span>","children":null,"spread":false},{"title":"Viewing the Iris Dataset with Pandas.ipynb <span style='color:#111;'> 20.57KB </span>","children":null,"spread":false},{"title":"Writing A Stacking Agreggator with Scikit-Learn Extra.ipynb <span style='color:#111;'> 20.36KB </span>","children":null,"spread":false},{"title":"Stacking with a Neural Network-checkpoint.ipynb <span style='color:#111;'> 19.85KB </span>","children":null,"spread":false},{"title":"Stacking with a Neural Network.ipynb <span style='color:#111;'> 19.85KB </span>","children":null,"spread":false},{"title":"Optimizing the number of centroids-checkpoint.ipynb <span style='color:#111;'> 19.77KB </span>","children":null,"spread":false},{"title":"Optimizing the number of centroids.ipynb <span style='color:#111;'> 19.77KB </span>","children":null,"spread":false},{"title":"ROC Analysis.ipynb <span style='color:#111;'> 16.22KB </span>","children":null,"spread":false},{"title":"Putting it All Together UCI Breast Cancer Data Set-checkpoint.ipynb <span style='color:#111;'> 15.29KB </span>","children":null,"spread":false},{"title":"Putting it All Together UCI Breast Cancer Data Set.ipynb <span style='color:#111;'> 15.29KB </span>","children":null,"spread":false},{"title":"Fitting a line through data-checkpoint.ipynb <span style='color:#111;'> 13.49KB </span>","children":null,"spread":false},{"title":"Fitting a line through data.ipynb <span style='color:#111;'> 13.49KB </span>","children":null,"spread":false},{"title":"Using truncated SVD to reduce dimensionality.ipynb <span style='color:#111;'> 11.77KB </span>","children":null,"spread":false},{"title":"Using stochastic gradient descent for regression.ipynb <span style='color:#111;'> 11.45KB </span>","children":null,"spread":false},{"title":"Create a Simple Estimator-checkpoint.ipynb <span style='color:#111;'> 11.44KB </span>","children":null,"spread":false},{"title":"Create a Simple Estimator.ipynb <span style='color:#111;'> 11.44KB </span>","children":null,"spread":false},{"title":"Numpy Basics.ipynb <span style='color:#111;'> 11.41KB </span>","children":null,"spread":false},{"title":"Viewing the Iris Dataset-checkpoint.ipynb <span style='color:#111;'> 11.02KB </span>","children":null,"spread":false},{"title":"Viewing the Iris Dataset.ipynb <span style='color:#111;'> 11.02KB </span>","children":null,"spread":false},{"title":"Feature selection.ipynb <span style='color:#111;'> 10.86KB </span>","children":null,"spread":false},{"title":"Fitting a line through data with Machine Learning-checkpoint.ipynb <span style='color:#111;'> 10.03KB </span>","children":null,"spread":false},{"title":"Fitting a line through data with Machine Learning.ipynb <span style='color:#111;'> 10.03KB </span>","children":null,"spread":false},{"title":"Classification Metrics-checkpoint.ipynb <span style='color:#111;'> 8.96KB </span>","children":null,"spread":false},{"title":"Classification Metrics.ipynb <span style='color:#111;'> 8.96KB </span>","children":null,"spread":false},{"title":"Perceptron Classifier-checkpoint.ipynb <span style='color:#111;'> 8.85KB </span>","children":null,"spread":false},{"title":"Perceptron Classifier.ipynb <span style='color:#111;'> 8.85KB </span>","children":null,"spread":false},{"title":"Plotting an ROC Curve without Context-checkpoint.ipynb <span style='color:#111;'> 8.23KB </span>","children":null,"spread":false},{"title":"Plotting an ROC Curve without Context.ipynb <span style='color:#111;'> 8.23KB </span>","children":null,"spread":false},{"title":"Classifying documents with Naive Bayes.ipynb <span style='color:#111;'> 7.97KB </span>","children":null,"spread":false},{"title":"Randomized Search with scikit-learn-checkpoint.ipynb <span style='color:#111;'> 7.73KB </span>","children":null,"spread":false},{"title":"Randomized Search with scikit-learn.ipynb <span style='color:#111;'> 7.73KB </span>","children":null,"spread":false},{"title":"Working with categorical variables-checkpoint.ipynb <span style='color:#111;'> 7.01KB </span>","children":null,"spread":false},{"title":"Working with categorical variables.ipynb <span style='color:#111;'> 7.01KB </span>","children":null,"spread":false},{"title":"Imputing missing values through various strategies-checkpoint.ipynb <span style='color:#111;'> 6.87KB </span>","children":null,"spread":false},{"title":"Imputing missing values through various strategies.ipynb <span style='color:#111;'> 6.87KB </span>","children":null,"spread":false},{"title":"Machine Learning with Logistic Regression.ipynb <span style='color:#111;'> 6.85KB </span>","children":null,"spread":false},{"title":"Regression Metrics-checkpoint.ipynb <span style='color:#111;'> 6.80KB </span>","children":null,"spread":false},{"title":"Regression Metrics.ipynb <span style='color:#111;'> 6.80KB </span>","children":null,"spread":false},{"title":"......","children":null,"spread":false},{"title":"<span style='color:steelblue;'>文件过多,未全部展示</span>","children":null,"spread":false}],"spread":true}]