Exploration of Biodegradable Substances Using Machine Learning Techniques
Abstract
:1. Introduction
- Data preprocessing and partitioning: Constant and nearly constant descriptors were eliminated. Correlated descriptors with a correlation coefficient exceeding 98% were represented by a single descriptor. The data records were then divided into training and test subsets.
- Feature ranking and selection: The remaining descriptors underwent feature ranking using minimum redundancy maximum relevance (mRMR) [22], a chi-square test (CHISQ) [23], and regularized neighborhood component analysis (RNCA) [24]. Each ranking method fed the most predictive features one by one into the three CART models, with cross-validation errors computed at each step. The feature subset with the minimum error was selected.
- Machine learning modeling: CARTs, SVM, kNN, and RLR models were built for biodegradation classification using the BO [25] and repeated cross-validation algorithms.
- In CART modeling, surrogate splits were employed to handle missing values, while other models processed imputed missing values [26].
- Experimental results demonstrated that the proposed CART model, incorporating curvature–interaction tests, achieved the highest performance in classifying the test subset. It achieved accuracy of 85.63%, sensitivity of 87.12%, specificity of 84.94%, and a highly comparable area under the ROC curve of 0.87. The model selected the top ten most important descriptors, with the SpMaxB(m) and SpMin1_Bh(v) descriptors significantly outperforming the others.
- A concise CART tree was constructed using these top ten features, yielding remarkable results with accuracy of 85.8%, sensitivity of 85.9%, and specificity of 85.8% for the test subset. The compact tree demonstrated explanatory transparency by providing predictive decision alternatives.
2. Literature Review
3. Method Pipeline
3.1. Data Collection and Preprocessing
3.2. Feature Selection
- The mRMR algorithm processes all features to find the optimal set that differs mutually and maximally and can effectively represent the target output. The algorithm quantifies the mutual information criterion to minimize the feature redundancy and maximize the relevance of the output.
- The CHISQ evaluates the individual chi-square test result (p-value) between each predictive feature and the output. The lower the p-value between the feature and the response, the higher the importance of the feature, and vice versa.
- RNCA leverages the Mahalanobis distance measure, commonly employed in kNN classification algorithms. The primary objective is to identify the most suitable subset of predictive features that maximizes the average leave-one-out classification accuracy over the training data. To mitigate overfitting, RNCA integrates a Gaussian prior into the neighborhood component analysis objective, resulting in a regularization method that greatly enhances the generalization capabilities.
3.3. Standard Classification and Regression Tree (CART)
- Gini index
- Deviance index
3.4. CART with Curvature and Interaction Tests
- Numeric features are partitioned into their quartiles. They are converted into the nominal type that bins record to the quartile according to their values. An extra bin is added for missing values (if they exist).
- For every class value in yk, k = 1,…, K, and every level in the partitioned feature xj, j = 1,…, J, the algorithm calculates the weighted number of records in class k as follows:
- If the p-value for the test <0.05, then the null hypothesis is rejected. The algorithm selects the splitting feature that minimizes the significant p-value (those less than 0.05). It is an unbiased selection regarding the number of levels in individual features, which provides a better interpretation of decision alternatives and a better ranking of predictive features according to their true importance. Curvature tests detect nonlinearities in the relationships between input features and the target variable and construct split points that capture the nonlinearity. This helps to improve the accuracy of predictions, particularly when the relationship between the features and target variable is complex.
- The interaction test is used to determine whether two features should be combined into a single predictor variable. The test minimizes the p-value of the chi-square tests of independence between every feature pair and the class variable. This test uses similar statistical procedures to evaluate the null hypothesis to assess the association between every pair of features for the target variable. In situations where there are several irrelevant features, interaction tests enable the identification of important features by examining the joint effect of two or more features on the target variable. Interaction tests, on the other hand, assist in identifying important features that may be overlooked by standard trees.
3.5. Bayesian Optimization (BO)
- It starts by sampling the true objective function at some random seed points to construct the initial dataset (D0). Then, the algorithm initializes the surrogate model SF0.
- At each iteration t, the AF finds the point that minimizes the SF model. This point represents the best guess to record the true objective function. The input point and the resulting function value update the dataset (Dt) and the SFt model.
- The algorithm reapplies the AF function to find the point that minimizes the updated SFt to estimate the new candidate point and so on.
- The iteration continues several times until satisfactory information is available about the objective function and then the global minimum is obtained.
3.5.1. Gaussian Process
3.5.2. Expected Improvement (EI)
4. Experimental Results
4.1. Feature Ranking
4.2. CARTs: Training and Evaluation
4.3. Model Comparisons
- Support vector machines (SVM) with radial basis function (RBF) kernels are widely used in machine learning for classification tasks [44]. The RBF kernel effectively separates classes in SVMs. Training an SVM with the RBF kernel requires consideration of two important parameters: C and gamma. Parameter C, common to all SVM kernels, controls the balance between training record misclassification and decision surface simplicity. A smaller C allows for a wider margin but may lead to more misclassifications, while a larger C aims to minimize misclassifications but may result in a narrower margin. The gamma parameter, specific to the RBF kernel, determines the influence of each training record on the decision boundary. A higher gamma value creates a more complex decision boundary, potentially causing overfitting, while a lower gamma value produces a smoother decision boundary, which may result in underfitting. The BO algorithm was employed to find the optimal values of C and gamma. Using the top mRMR 60 features, the study achieved general accuracies of 89.19% for the training subset and 83.30% for the test subset. In contrast, utilizing only the top 10 features recommended by the CART models resulted in accuracies of 86.94% for the training subset and 82.14% for the test subset, as shown in Table 4.
- The K-nearest neighbors (kNN) algorithm is a popular choice in solving classification problems in machine learning [45]. It is a non-parametric, supervised learning classifier that relies on closely related features to make predictions or classifications for individual data points. In kNN, classification is based on the idea that data records with closely related features are likely to belong to the same class. The algorithm identifies the K-nearest neighbors of a given record in the feature space and assigns the class label based on a majority vote among these neighbors. The choice of K (the number of neighbors) and the distance function are crucial hyperparameters that can be tuned to optimize performance. In this study, the BO algorithm was used to determine the optimal values of K and the distance function. Using the top 60 mRMR features, the kNN model achieved accuracies of 100.0% for the training subset and 83.5% for the test subset. Similarly, when considering only the top ten features recommended by the CART models, the accuracies were reported as 100.0% for the training subset and 82.52% for the test subset, as detailed in Table 4.
- Logistic regression (LR) is a commonly used classification algorithm that models the relationships between input variables and a binary outcome using a logistic function [46]. The logistic function produces an S-shaped curve that maps inputs to a probability value between 0 and 1, representing the predicted probability of a positive outcome. The model estimates the logistic function’s parameters using maximum likelihood estimation. Regulated LR (RLR) utilizes regularization to prevent overfitting and improve generalization by adding a penalty term to the cost function. This penalty term reduces the magnitude of coefficients and prevents them from growing too large. Two popular regularization techniques in logistic regression are L1 (lasso) and L2 (ridge) regularization. L1 regularization adds the absolute values of coefficients to the cost function, causing some coefficients to become exactly zero. L2 regularization adds the squared values of coefficients to the cost function. In this study, the BO algorithm was used to find the optimal values for regularization strength (lambda) and regularization penalty type (L1 or L2). When utilizing the top 60 mRMR-selected features, the logistic regression model achieved accuracies of 78.04% for the training subset and 74.18% for the test subset. However, when considering only the top ten features recommended by the CART models, the model’s performance resulted in accuracies of 75.96% for the training subset and 074.56% for the test subset, as outlined in Table 4.
4.4. Model Evaluation Using ROC Curves
5. Conclusions and Discussion
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
References
- Gu, J.D. Biodegradability of plastics: The issues, recent advances, and future perspectives. Environ. Sci. Pollut. Res. 2021, 28, 1278–1282. [Google Scholar] [CrossRef] [PubMed]
- Pagga, U. Testing biodegradability with standardized methods. Chemosphere 1997, 35, 2953–2972. [Google Scholar] [CrossRef] [PubMed]
- Grisoni, F.; Ballabio, D.; Todeschini, R.; Consonni, V. Molecular Descriptors for Structure–Activity Applications: A Hands-On Approach. In Computational Toxicology: Methods and Protocols; Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–53. [Google Scholar]
- Anastas, P.T.; Kirchhoff, M.M. Origins, Current Status, and Future Challenges of Green Chemistry. Acc. Chem. Res. 2002, 35, 686–694. [Google Scholar] [CrossRef] [PubMed]
- Villaverde, J.; Sevilla-Morán, B.; López-Goti, C.; Alonso-Prados, J.; Sandín-España, P. QSAR/QSPR models based on quantum chemistry for risk assessment of pesticides according to current European legislation. SAR QSAR Environ. Res. 2019, 31, 49–72. [Google Scholar] [CrossRef]
- Kazue, C.; Malloy, T. QSAR Use in REACH analyses of alternatives to predict human health and environmental toxicity of alternative chemical substances. Integr. Environ. Assess. Manag. 2020, 16, 745–760. [Google Scholar]
- Czermiński, R.; Abdelaziz, Y.; Hartsough, D. Use of support vector machine in pattern classification: Application to QSAR studies. Quant. Struct. Act. Relatsh. 2001, 20, 227–240. [Google Scholar] [CrossRef]
- Ghasemi, F.; Mehridehnavi, A.; Pérez-Garrido, A.; Pérez-Sánchez, H. Neural network and deep-learning algorithms used in QSAR studies: Merits and drawbacks. Drug Discov. Today 2018, 23, 1784–1790. [Google Scholar] [CrossRef]
- Rocha, W.F.C.; Sheen, D.A. Classification of biodegradable materials using QSAR modelling with uncertainty estimation. SAR QSAR Environ. Res. 2016, 27, 799–811. [Google Scholar] [CrossRef]
- Ajmani, S.; Jadhav, K.; Kulkarni, S.A. Three-Dimensional QSAR Using the k-Nearest Neighbor Method and Its Interpretation. J. Chem. Inf. Model. 2005, 46, 24–31. [Google Scholar] [CrossRef]
- Kotsiantis, S.B. Decision trees: A recent overview. Artif. Intell. Rev. 2013, 39, 261–283. [Google Scholar] [CrossRef]
- Mienye, I.B.; Sun, Y.; Wang, Z. Prediction performance of improved decision tree-based algorithms: A review. Procedia Manuf. 2019, 35, 698–703. [Google Scholar] [CrossRef]
- Podgorelec, V.; Kokol, P.; Stiglic, B.; Rozman, I. Decision Trees: An Overview and Their Use in Medicine. J. Med. Syst. 2002, 26, 445–463. [Google Scholar] [CrossRef]
- Dudkina, T.; Meniailov, I.; Bazilevych, K.; Krivtsov, S.; Tkachenko, A. Classification and Prediction of Diabetes Disease using Decision Tree Method. In Proceedings of the IT&AS 2021: Symposium on Information Technologies & Applied Sciences, Bratislava, Slovakia, 5 March 2021; pp. 163–172. [Google Scholar]
- Koteluk, O.; Wartecki, A.; Mazurek, S.; Kołodziejczak, I.; Mackiewicz, A. How Do Machines Learn? Artificial Intelligence as a New Era in Medicine. J. Pers. Med. 2021, 11, 32. [Google Scholar]
- Breiman, L. Classification and Regression Trees; Routledge: New York, NY, USA, 2017. [Google Scholar]
- Bühlmann, P. Bagging, boosting and ensemble methods. In Handbook of Computational Statistics: Concepts and Methods; Springer: Berlin/Heidelberg, Germany, 2012; pp. 985–1022. [Google Scholar]
- Loh, W.Y.; Shih, Y.S. Split selection methods for classification trees. Stat. Sin. 1997, 7, 815–840. [Google Scholar]
- Mansouri, K.; Ringsted, T.; Ballabio, D.; Todeschini, R.; Consonni, V. Quantitative Structure–Activity Relationship Models for Ready Biodegradability of Chemicals. J. Chem. Inf. Model. 2013, 53, 867–878. [Google Scholar] [CrossRef] [PubMed]
- Jacobs, A.; Williams, D.; Hickey, K.; Patrick, N.; Williams, A.J.; Chalk, S.; McEwen, L.; Willighagen, E.; Walker, M.; Bolton, E.; et al. CAS Common Chemistry in 2021: Expanding Access to Trusted Chemical Information for the Scientific Community. J. Chem. Inf. Model. 2022, 62, 2737–2743. [Google Scholar] [CrossRef]
- Mauri, A.; Bertola, M. Alvascience: A New Software Suite for the QSAR Workflow Applied to the Blood–Brain Barrier Permeability. Int. J. Mol. Sci. 2022, 23, 12882. [Google Scholar] [CrossRef] [PubMed]
- Bugata, P.; Drotar, P. On some aspects of minimum redundancy maximum relevance feature selection. Sci. China Inf. Sci. 2019, 63, 112103. [Google Scholar] [CrossRef]
- Thaseen, I.S.; Cherukuri, A.K. Intrusion detection model using fusion of chi-square feature selection and multi class SVM. J. King Saud Univ. Comput. Inf. Sci. 2017, 29, 462–472. [Google Scholar]
- Yang, W.; Wang, K.; Zuo, W. Neighborhood Component Feature Selection for High-Dimensional Data. J. Comput. 2012, 7, 161–168. [Google Scholar] [CrossRef]
- Dewancker, I.; McCourt, M.; Clark, S. Bayesian optimization for machine learning: A practical guidebook. arXiv 2016, arXiv:1612.04858. [Google Scholar]
- Thirukumaran, S.; Sumathi, A. Missing value imputation techniques depth survey and an imputation algorithm to improve the efficiency of imputation. In Proceedings of the 2012 Fourth International Conference on Advanced Computing (ICoAC), Chennai, India, 13–15 December 2012; IEEE: Piscataway Township, NJ, USA; pp. 1–5. [Google Scholar]
- Singh, A.K.; Bilal, M.; Iqbal, H.M.; Raj, A. Trends in predictive biodegradation for sustainable mitigation of environmental pollutants: Recent progress and future outlook. Sci. Total Environ. 2021, 770, 144561. [Google Scholar] [CrossRef]
- Lee, M.; Min, K. A comparative study of the performance for predicting biodegradability classification: The quantitative structure–activity relationship model vs the graph convolutional network. ACS Omega 2022, 7, 3649–3655. [Google Scholar] [CrossRef] [PubMed]
- Cereto-Massagué, A.; Ojeda, M.J.; Valls, C.; Mulero, M.; Garcia-Vallvé, S.; Pujadas, G. Molecular fingerprint similarity search in virtual screening. Methods 2015, 71, 58–63. [Google Scholar] [CrossRef] [PubMed]
- Silva, G.M.; Federico, L.B.; Alves, V.M.; Silva, C.H.T. In silico methods to predict relevant toxicological endpoints of bioactive substances. In Functional Properties of Advanced Engineering Materials and Biomolecules; Springer International Publishing: Cham, Switzerland, 2021; pp. 649–676. [Google Scholar]
- Gu, W.; Li, Q.; Li, Y. Law and mechanism analysis of biodegradability of polychlorinated naphthalenes based on principal component analysis, QSAR models, molecular docking and molecular dynamics simulation. Chemosphere 2020, 243, 125427. [Google Scholar] [CrossRef]
- Lunghini, F.; Marcou, G.; Gantzer, P.; Azam, P.; Horvath, D.; Van Miert, E.; Varnek, A. Modelling of ready biodegradability based on combined public and industrial data sources. SAR QSAR Environ. Res. 2020, 31, 171–186. [Google Scholar] [CrossRef]
- Putra, R.I.D.; Maulana, A.L.; Saputro, A.G. Study on building machine learning model to predict biodegradable-ready materials. AIP Conf. Proc. 2019, 2088, 060003. [Google Scholar]
- Elsayad, A.M.; Ahmed, M.; Al-Dhaifallah, N.M.; Khaled, A.E. Classification of biodegradable substances using balanced random trees and boosted C5. 0 Decision Trees. Int. J. Environ. Res. Public Health 2020, 17, 9322. [Google Scholar] [CrossRef]
- Marín-Ortega, P.M.; Dmitriyev, V.; Abilov, M.; Gómez, J.M. ELTA: New approach in designing business intelligence solutions in era of big data. Procedia Technol. 2014, 16, 667–674. [Google Scholar] [CrossRef]
- Li, X.; Yin, B.; Tian, W.; Sun, Y. Performance of repeated cross validation for machine learning models in building energy analysis. In Proceedings of the 11th International Symposium on Heating, Ventilation and Air Conditioning (ISHVAC 2019) Volume III: Buildings and Energy 11, Harbin, China, 12–15 July 2019; Springer: Singapore, 2020; pp. 523–531. [Google Scholar]
- Frazier, P.I. A tutorial on Bayesian optimization. arXiv 2018, arXiv:1807.02811. [Google Scholar]
- Xu, Z.; Guo, Y.; Saleh, J.H. Efficient hybrid Bayesian optimization algorithm with adaptive expected improvement acquisition function. Eng. Optim. 2021, 53, 1786–1804. [Google Scholar] [CrossRef]
- Christianson, R.B.; Gramacy, R.B. Robust expected improvement for Bayesian optimization. arXiv 2023, arXiv:2302.08612. [Google Scholar]
- Wen, Z.; Nancy Zeng, N.; Wang, N. Sensitivity, specificity, accuracy, associated confidence interval and ROC analysis with practical SAS implementations. In Proceedings of the Northeast SAS Users Group (NESUG) conference: Health care and life sciences, Baltimore, MD, USA, 4–17 November 2010; Volume 19, p. 67. [Google Scholar]
- Kalina, J.; Schlenker, A. A robust supervised variable selection for noisy high-dimensional data. BioMed Res. Int. 2015, 2015, 320385. [Google Scholar] [CrossRef] [PubMed]
- Očenášek, J.; Schwarz, J. The parallel Bayesian optimization algorithm. In Proceedings of the State of the Art in Computational Intelligence: Proceedings of the European Symposium on Computational Intelligence, Košice, Slovakia, 30 August–1 September 2000; pp. 61–67. [Google Scholar]
- Krzywinski, M.; Naomi Altman, N. Classification and regression trees. Nat. Methods 2017, 14, 757–758. [Google Scholar] [CrossRef]
- Zhang, Y. Support vector machine classification algorithm and its application. In Proceedings of the Information Computing and Applications: Third International Conference, ICICA 2012, Chengde, China, 14–16 September 2012; Proceedings, Part II 3. Springer: Berlin/Heidelberg, Germany, 2012; pp. 179–186. [Google Scholar]
- Taunk, K.; De, S.; Verma, V.; Swetapadma, A. A brief review of nearest neighbor algorithm for learning and classification. In Proceedings of the 2019 International Conference on Intelligent Computing and Control Systems (ICCS), Madurai, India, 15–17 May 2019; IEEE: Piscataway Township, NJ, USA, 2019; pp. 1255–1260. [Google Scholar]
- Li, W.; Liu, H.; Yang, P.; Wei, X. Supporting regularized logistic regression privately and efficiently. PLoS ONE 2016, 11, e0156479. [Google Scholar] [CrossRef] [PubMed]
Model | CV Error | Training | Testing | Model Parameters | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|
Acc | Sen | Sps | Acc | Sen | Sps | Split Criterion | Min Parent | Min Leaf | Max Splits | ||
All Splits | 0.18 | 85.36 | 86.65 | 84.76 | 84.66 | 82.82 | 85.51 | deviance | 10 | 2 | 34 |
Curvature | 0.17 | 86.69 | 92.67 | 83.90 | 81.55 | 85.89 | 79.55 | deviance | 10 | 1 | 59 |
Curvature–Interaction | 0.17 | 85.77 | 89.53 | 84.02 | 85.63 | 87.12 | 84.94 | deviance | 10 | 2 | 35 |
Range | Minimum | Maximum | Mean | Std Deviation | Skewness | Kurtosis | |
---|---|---|---|---|---|---|---|
SpMax_B_m | 15.4 | 2.19 | 17.58 | 3.9969 | 1.16114 | 4.318 | 27.649 |
SpMin1_Bh_v | 2.33 | 0 | 2.33 | 1.9776 | 0.13717 | −3.864 | 39.381 |
DLS_02 | 0.83 | 0.17 | 1 | 0.7672 | 0.16699 | −0.461 | −0.275 |
N_069 | 3 | 0 | 3 | 0.0957 | 0.35575 | 4.081 | 17.589 |
B05_N_N | 1 | 0 | 1 | 0.074 | 0.26195 | 3.258 | 8.626 |
B03_N_N | 1 | 0 | 1 | 0.0724 | 0.25922 | 3.305 | 8.936 |
NsssN | 4 | 0 | 4 | 0.1622 | 0.49423 | 3.849 | 17.529 |
nCq | 9 | 0 | 9 | 0.1015 | 0.51264 | 8.585 | 105.722 |
rNHR | 2 | 0 | 2 | 0.0549 | 0.2891 | 5.636 | 32.251 |
B05_N_Cl | 1 | 0 | 1 | 0.0333 | 0.17944 | 5.211 | 25.194 |
Model | cv Error | Training | Testing | Model Parameters | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|
Acc | Sen | Sps | Acc | Sen | Sps | Split Criterion | Min Parent | Min Leaf | Max Splits | ||
All Splits | 0.1727 | 84.9 | 86.4 | 84.3 | 84.9 | 84.7 | 84.9 | gdi | 10 | 1 | 29 |
Curvature | 0.1762 | 84.8 | 87.7 | 83.4 | 84.7 | 86.5 | 83.8 | ‘deviance’ | 10 | 1 | 32 |
Curvature–Interaction | 0.1742 | 84.8 | 86.6 | 83.9 | 85.8 | 85.9 | 85.8 | ‘deviance’ | 10 | 5 | 25 |
Model | CV Error | Training | Testing | |||||
---|---|---|---|---|---|---|---|---|
Acc | Sen | Sps | Acc | Sen | Sps | |||
60 mRMR features | SVM | 0.15 | 89.19 | 93.46 | 87.20 | 83.30 | 83.44 | 83.24 |
kNN | 0.15 | 100.00 | 100.00 | 100.00 | 83.50 | 81.60 | 84.38 | |
RLR | 0.19 | 78.04 | 92.41 | 71.34 | 74.18 | 88.34 | 67.61 | |
Top Ten Features | SVM | 0.16 | 86.94 | 89.79 | 85.61 | 82.14 | 80.98 | 82.67 |
kNN | 0.16 | 100.00 | 100.00 | 100.00 | 82.52 | 83.44 | 82.10 | |
RLR | 0.00 | 75.96 | 89.53 | 69.63 | 74.56 | 87.12 | 68.75 |
Model/AUROC | Training Subset | Test Subset |
---|---|---|
CART All Splits | 0.90 | 0.88 |
CART Curvature | 0.93 | 0.84 |
CART Curvature–Interaction | 0.90 | 0.87 |
SVM | 0.96 | 0.86 |
kNN | 1 | 0.89 |
DA | 0.90 | 0.88 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Elsayad, A.M.; Zeghid, M.; Ahmed, H.Y.; Elsayad, K.A. Exploration of Biodegradable Substances Using Machine Learning Techniques. Sustainability 2023, 15, 12764. https://doi.org/10.3390/su151712764
Elsayad AM, Zeghid M, Ahmed HY, Elsayad KA. Exploration of Biodegradable Substances Using Machine Learning Techniques. Sustainability. 2023; 15(17):12764. https://doi.org/10.3390/su151712764
Chicago/Turabian StyleElsayad, Alaa M., Medien Zeghid, Hassan Yousif Ahmed, and Khaled A. Elsayad. 2023. "Exploration of Biodegradable Substances Using Machine Learning Techniques" Sustainability 15, no. 17: 12764. https://doi.org/10.3390/su151712764
APA StyleElsayad, A. M., Zeghid, M., Ahmed, H. Y., & Elsayad, K. A. (2023). Exploration of Biodegradable Substances Using Machine Learning Techniques. Sustainability, 15(17), 12764. https://doi.org/10.3390/su151712764