Today, I finished the NMR paper.
Summary:
I sent you by an email.
I did three tests today for selecting inputs:
1. delete 10 inversion logs: R2 dropped to 0.7521.
2. delete 12 conventional logs: R2 dropped to 0.7889.
3. delete AT90 and VPVS logs: R2 dropped to 0.8423.
For the first two cases, apparent drops are obtained. So I recommend not to delete either of them. For the third case, the least important two inputs are deleted and similar accuracy as the original one is obtained. But since we have analyzed the importance of inputs in the paper, I think there is no need to delete them in this paper. Actually, I remembered that we discussed about it before. First, there is a big accuracy drop without either conventional logs or inversion logs. Second, since we have all of these 22 logs (without flags), it is better to use them together to predict NMR t2 distribution. There is no need to delete 2 or 3 of them if we want to keep similar accuracy.
I did not use k-fold cross-validation method because 85% training data and 15% testing data have been selected already before training and testing models. Instead, I did parallel computing for dividing data and training models. Almost the same accuracy is obtained so that I did not change results and figures in the paper.
The changed paper can be seen in the email.
5/08/2018
NMR sequential prediction
-
column d is prediction of every group in one ANN model separately.
-
column e is sequential prediction.
Column a is accuracy of predicting 64 bins together and measuring accuracy of every group while normalized using maximum and minimum values of 64 bins. Column b is the ranking result. Column c is the number of bins
in every group.
From the results, we can see that the two methods that we discussed yesterday do not perform better than predicting 64 bins together. The reason may be:
-
64 bins together represent the distribution of pore size.
-
some group just means some pore size, which cannot be closely correlated to those input logs. The concentration of pore size at every depth is different from each other.
-
what matters is the dominated pore size at every depth, which is different from each other.
I think the sequential prediction for NMR and DD logs are different. 8 DD logs are kind of independent of each other at different frequencies, each representing conductivity or permittivity of the reservoir. However,
64 bins and 8 groups at every depth represent pore size distribution together, which cannot be divided. I think that is why the two methods cannot predict NMR t2 distribution separately with high accuracy.
Tomorrow, I will select inputs and do k-fold cross validation.
4/24/2018
prepare for the defense
Today, I prepared for the defense.
Summary:
I recited the contents for slides. I also changed some slides from contents to figures.
Tomorrow, I will continue to prepare for the defense.
Summary:
I recited the contents for slides. I also changed some slides from contents to figures.
Tomorrow, I will continue to prepare for the defense.
4/19/2018
finish the ppt
Today, I finish the ppt for defense.
Summary:
I prepare for contents in every slide.
Tomorrow, I will deal with your suggestions for NMR paper.
Summary:
I prepare for contents in every slide.
Tomorrow, I will deal with your suggestions for NMR paper.
4/18/2018
prepare for the presentation
Today, I prepare for the presentation.
Summary:
I think about what to say in every slide.
Tomorrow, I will continue to prepare for it.
Summary:
I think about what to say in every slide.
Tomorrow, I will continue to prepare for it.
4/17/2018
finish the draft of ppt
Today, I finish the draft of ppt.
Summary:
I just finish all contents in the ppt.
Tomorrow, I will start to prepare the presentation and improve the ppt.
Summary:
I just finish all contents in the ppt.
Tomorrow, I will start to prepare the presentation and improve the ppt.
4/16/2018
make ppt for defense
Today, I made PPT for defense.
Summary:
I finish about 50%.
Tomorrow, I will try to finish all of them.
Summary:
I finish about 50%.
Tomorrow, I will try to finish all of them.
4/12/2018
improve the thesis
Today, I continue to improve the thesis.
Summary:
I may finish about 70%.
Tomorrow, I will try to finish the thesis.
Summary:
I may finish about 70%.
Tomorrow, I will try to finish the thesis.
4/11/2018
improve the paper
Today, I continue to improve the paper.
Summary:
I may finish about 50%.
Tomorrow, I will continue to improve the paper.
Summary:
I may finish about 50%.
Tomorrow, I will continue to improve the paper.
4/10/2018
improve the second version of thesis
Today, I started to improve the second version of thesis.
Summary:
I may finish 20%.
Tomorrow, I will continue to improve it.
Summary:
I may finish 20%.
Tomorrow, I will continue to improve it.
4/09/2018
k-mean method
Today, I compared clustering methods.
Summary:
I compared those clustering methods before and decided to use k-mean method to do clustering.
I am now looking for the best method to optimize the number of clusters.
Tomorrow, I will continue to do research on it.
Summary:
I compared those clustering methods before and decided to use k-mean method to do clustering.
I am now looking for the best method to optimize the number of clusters.
Tomorrow, I will continue to do research on it.
3/29/2018
select figures
Today, I selected figures for defense.
Summary:
I sent you by an email.
Tomorrow, I will start to write the thesis.
Summary:
I sent you by an email.
Tomorrow, I will start to write the thesis.
3/28/2018
analyze 15 logs
Today, I analyze 15 logs
Summary:
The result is also not good for all models.
Tomorrow, I will try different number of logs and start to write the thesis.
Summary:
The result is also not good for all models.
Tomorrow, I will try different number of logs and start to write the thesis.
3/27/2018
add SOM
Today, I added SOM.
Summary:
For now, I just cluster sonic log, namely DTC and DTS because it is visible. From the scatter plot, we can see that sonic log is better to be divided into 2 or 3 clusters, which are close to prediction performance levels. As a result, the second and third figures are plotted after comparing prediction performance with 4 clustering models. However, some of them are close to prediction results while others are not.
I think that is the best that I can do if only sonic log is included for clustering.
Sonic plot:
Two clusters:
Three clusters:
Tomorrow, I will include some inputs into these models to see if more clusters can be determined and if results could be improved.
Summary:
For now, I just cluster sonic log, namely DTC and DTS because it is visible. From the scatter plot, we can see that sonic log is better to be divided into 2 or 3 clusters, which are close to prediction performance levels. As a result, the second and third figures are plotted after comparing prediction performance with 4 clustering models. However, some of them are close to prediction results while others are not.
I think that is the best that I can do if only sonic log is included for clustering.
Sonic plot:
Two clusters:
Three clusters:
Tomorrow, I will include some inputs into these models to see if more clusters can be determined and if results could be improved.
3/26/2018
clustering models
Today, I built three clustering models for sonic log.
Summary:
Three models are:
1. k-means clustering
2. density-based spatial clustering of applications with noise (DBSCAN)
3. expectation maximization algorithm for gaussian mixture model (EM GMM)
The following are results of two clusters and three clusters.
Two clusters
Three clusters
From two figures, I think that two clusters performs better than three clusters. For k-means, it includes medium and poor clusters together and for DBSCAN and EM GMM, they include good and medium clusters together.
Tomorrow, I will try to improve the results.
Summary:
Three models are:
1. k-means clustering
2. density-based spatial clustering of applications with noise (DBSCAN)
3. expectation maximization algorithm for gaussian mixture model (EM GMM)
The following are results of two clusters and three clusters.
Two clusters
Three clusters
From two figures, I think that two clusters performs better than three clusters. For k-means, it includes medium and poor clusters together and for DBSCAN and EM GMM, they include good and medium clusters together.
Tomorrow, I will try to improve the results.
2/26/2018
write the outline of sonic log paper
Today, I wrote the outline of sonic log paper.
Summary:
I will finish it when it is done today.
Tomorrow, I will start to write the paper.
Summary:
I will finish it when it is done today.
Tomorrow, I will start to write the paper.
2/23/2018
finish the Geophysics paper and start the sonic paper
Today, I finished the Geophysics paper and started the sonic paper.
Summary:
I sent you by an email.
Next week, I will continue to work on the sonic paper.
Summary:
I sent you by an email.
Next week, I will continue to work on the sonic paper.
2/22/2018
change the comments
Today, I changed the comments.
Summary:
I am now writing the modification part.
Tomorrow, I will finish it.
Summary:
I am now writing the modification part.
Tomorrow, I will finish it.
2/21/2018
finish the improvement
Today, I finished the improvement of the geophysics paper.
Summary:
I sent you by an email.
Tomorrow, I will improve some contents of the paper.
Summary:
I sent you by an email.
Tomorrow, I will improve some contents of the paper.
2/19/2018
continue to improve the paper
Today, I continued to improve the paper.
Summary:
I cannot finish it today. It may take another 2 or 3 days. I will try to finish it this week.
Tomorrow, I will continue to improve the paper.
Summary:
I cannot finish it today. It may take another 2 or 3 days. I will try to finish it this week.
Tomorrow, I will continue to improve the paper.
2/16/2018
finish the GJI paper
Today, I finished changing the format for GJI paper.
Summary:
I changed the format for GJI paper and continue to improve the geophysics paper.
Next week, I will continue to improve the geophysics paper.
Summary:
I changed the format for GJI paper and continue to improve the geophysics paper.
Next week, I will continue to improve the geophysics paper.
2/15/2018
continue to improve the paper
Today, I continued to improve the paper.
Summary:
I go on with another 3 comments.
Tomorrow, I will change the format of the rejected paper.
Summary:
I go on with another 3 comments.
Tomorrow, I will change the format of the rejected paper.
2/14/2018
deal with the comments of the second reviewer
Today, I dealt with the comments of the second reviewer.
Summary:
There are about 20 comments from the second reviewer. I will deal with them one by one.
Tomorrow, I will continue to deal with them.
Summary:
There are about 20 comments from the second reviewer. I will deal with them one by one.
Tomorrow, I will continue to deal with them.
2/13/2018
improve the geophysics paper
Today, I continued to improve the geophysics paper.
Summary:
I have finished answering comments from the first reviewer.
Tomorrow: I will start to deal with the second reviewer.
Summary:
I have finished answering comments from the first reviewer.
Tomorrow: I will start to deal with the second reviewer.
2/12/2018
start to improve the geophysics paper
Today, I started to improve the geophysics paper.
Summary:
I marked all comments in four colors.
I started to deal with them one by one.
I will discuss with you about them if I have problems.
Tomorrow: I will continue to improve the paper.
Summary:
I marked all comments in four colors.
I started to deal with them one by one.
I will discuss with you about them if I have problems.
Tomorrow: I will continue to improve the paper.
2/09/2018
finish the NMR paper
Today, I finished the NMR paper.
Summary:
1. I read materials of feature selection. There should be no problem of my method to evaluate the importance of inputs in the ANN model.
2. There are many models to rank the importance or select a subset of input data. However, they are all targeted for specific models. For example, SVM-RFE is for SVM model. Relief-F is for regression model. As a result, ranking importance using these models will not show the actual importance in the ANN model.
3. So I recommend to rank importance in the ANN model by using the original method.
Next week, I may start for the geophysics paper.
Summary:
1. I read materials of feature selection. There should be no problem of my method to evaluate the importance of inputs in the ANN model.
2. There are many models to rank the importance or select a subset of input data. However, they are all targeted for specific models. For example, SVM-RFE is for SVM model. Relief-F is for regression model. As a result, ranking importance using these models will not show the actual importance in the ANN model.
3. So I recommend to rank importance in the ANN model by using the original method.
Next week, I may start for the geophysics paper.
2/08/2018
change the wrong figure
Today, I changed the wrong figure.
Summary:
The day before yesterday, I plotted the importance result. But there are two problems.
1. I deleted inputs one by one without using 0 to replace them.
2. The accuracy drop was not divided by original accuracy.
I fixed them and plot the following figure, which should be no problem.
Tomorrow, I will continue to look for methods to rank importance and give you the final version of the paper.
Summary:
The day before yesterday, I plotted the importance result. But there are two problems.
1. I deleted inputs one by one without using 0 to replace them.
2. The accuracy drop was not divided by original accuracy.
I fixed them and plot the following figure, which should be no problem.
Tomorrow, I will continue to look for methods to rank importance and give you the final version of the paper.
2/07/2018
rank importance of inputs
Today, I used other ways to rank importance of inputs.
Summary:
Feature selection is one method to decrease dimension of inputs while keeping as many information of inputs as possible. It select a subset of input for model construction. As a result, it is not for ranking importance of inputs.
I found another algorithm for ranking importance, which is called ReliefF.
The following is the result:
The ranking is different from what is calculated by deleting inputs one by one. As searched online, most people analyze importance by deleting inputs one by one. So I think that the original method is more trustworthy than the ReliefF algorithm.
Tomorrow, we can discuss about the result and I plan to go on the next task, which is find clustering method for sonic log prediction model.
Summary:
Feature selection is one method to decrease dimension of inputs while keeping as many information of inputs as possible. It select a subset of input for model construction. As a result, it is not for ranking importance of inputs.
I found another algorithm for ranking importance, which is called ReliefF.
The following is the result:
Tomorrow, we can discuss about the result and I plan to go on the next task, which is find clustering method for sonic log prediction model.
2/06/2018
finish accuracy drop analysis
Today, I finished accuracy drop analysis.
Summary:
The following is the result.
Tomorrow, I will try to use other ways to rank the importance of inputs.
Summary:
The following is the result.
Tomorrow, I will try to use other ways to rank the importance of inputs.
2/05/2018
coding for ranking inputs
Today, I coded for ranking inputs.
Summary:
I have not finished it yet. I will try to finish it tomorrow.
But my concern is that: since I have 27 inputs, deleting any one of them will not see apparent accuracy drop. So the ranking may not be feasible.
Tomorrow, I will try to run and see the results.
Summary:
I have not finished it yet. I will try to finish it tomorrow.
But my concern is that: since I have 27 inputs, deleting any one of them will not see apparent accuracy drop. So the ranking may not be feasible.
Tomorrow, I will try to run and see the results.
2/01/2018
evaluation of GMM application in accuracy analysis
Today, I did evaluation of GMM application in accuracy analysis.
Summary:
I have not coded for the whole process, so I did not know whether it will work. But we can have a try if you want.
Tomorrow: I will continue to read some materials about GMM.
Summary:
I have not coded for the whole process, so I did not know whether it will work. But we can have a try if you want.
Tomorrow: I will continue to read some materials about GMM.
1/31/2018
GMM may be for fitting problems.
Today, I learned about how to build a GMM in mATLAB.
Summary:
I knew how to use the GMM to fit data, but it may not be able for prediction.
GMM is used to fit data using several gaussian distributions. People can obtain parameters of these gaussian distributions from raw data, but it is about the fitting problem.
For now, I did not see any application of GMM similar to our problem.
Tomorrow, maybe we can have a talk about GMM.
Summary:
I knew how to use the GMM to fit data, but it may not be able for prediction.
GMM is used to fit data using several gaussian distributions. People can obtain parameters of these gaussian distributions from raw data, but it is about the fitting problem.
For now, I did not see any application of GMM similar to our problem.
Tomorrow, maybe we can have a talk about GMM.
1/30/2018
continue to learn about the model
Today, I continued to learn about the model.
Summary:
I am learning how to build the model in Matlab.
Tomorrow, I will continue to build the model.
Summary:
I am learning how to build the model in Matlab.
Tomorrow, I will continue to build the model.
1/29/2018
develop the gaussian mixture model
Today, I learned about the gaussian mixture model.
Summary:
I started to build the gaussian mixture model.
Tomorrow, I will continue to develop the model.
Summary:
I started to build the gaussian mixture model.
Tomorrow, I will continue to develop the model.
1/26/2018
finish improving ATCE paper proposal
Today, I finished improving ATCE paper proposal.
Summary:
I sent you by an email.
Next week, I will start to work on sonic log paper.
Summary:
I sent you by an email.
Next week, I will start to work on sonic log paper.
1/25/2018
improve two abstracts
Today, I improved two abstracts.
Summary:
I am now working on the ATCE paper proposal and SPWLA abstract.
Tomorrow, I will try to finish them.
Summary:
I am now working on the ATCE paper proposal and SPWLA abstract.
Tomorrow, I will try to finish them.
1/24/2018
finish the TGRS paper and start to write ATCE paper proposal
Today, I finished the TGRS paper and started to write ATCE paper proposal.
Summary:
I am now writing the ATCE paper proposal on sonic logs.
Tomorrow, I will try to finish the paper proposal.
Summary:
I am now writing the ATCE paper proposal on sonic logs.
Tomorrow, I will try to finish the paper proposal.
1/23/2018
continue to improve the paper
Today, I continued to improve the TGRS paper.
Summary:
I am now working on improving the sensitivity to noise in data.
Tomorrow, I will try to finish the TGRS paper.
Summary:
I am now working on improving the sensitivity to noise in data.
Tomorrow, I will try to finish the TGRS paper.
1/22/2018
continue to improve the TGRS paper
Today, I continued to improve the TGRS paper.
Summary:
I am now coding for re-ranking for all inputs to decide the number of inputs kept when accuracy drop does not exceed 10%.
Tomorrow, I will continue to improve the TGRS paper.
Summary:
I am now coding for re-ranking for all inputs to decide the number of inputs kept when accuracy drop does not exceed 10%.
Tomorrow, I will continue to improve the TGRS paper.
1/19/2018
continue to improve the TGRS paper
Today, I continued to improve the TGRS paper.
1.22-1.26
finish the TGRS paper, submit the abstract for SPWLA
1.29-2.2
build the gaussian mixture model and compare it with other models
2.5-2.9
test the best model in another well
2.12-2.16
write the draft of sonic log paper
2.19-2.23
improve and finish the sonic log paper
Next week, I will continue to improve the TGRS paper.
1.22-1.26
finish the TGRS paper, submit the abstract for SPWLA
1.29-2.2
build the gaussian mixture model and compare it with other models
2.5-2.9
test the best model in another well
2.12-2.16
write the draft of sonic log paper
2.19-2.23
improve and finish the sonic log paper
Next week, I will continue to improve the TGRS paper.
1/18/2018
finish the PPT and start to work on TGRS paper
Today, I finished the PPT and started to work in TGRS paper.
Summary:
I will send the PPT to Dr. Deepak and Dr. Wu tomorrow morning.
I will improve the abstract for SPWLA after you review it.
Tomorrow, I will continue to work on the TGRS paper.
Summary:
I will send the PPT to Dr. Deepak and Dr. Wu tomorrow morning.
I will improve the abstract for SPWLA after you review it.
Tomorrow, I will continue to work on the TGRS paper.
1/17/2018
finish the abstract for SPWLA
Today, I finished the abstract for SPWLA.
Summary:
I sent you by an email.
Tomorrow, I will make the ppt and improve the abstract.
Summary:
I sent you by an email.
Tomorrow, I will make the ppt and improve the abstract.
1/16/2018
finish checking the innovation and start to write the abstract
Today, I finished checking the innovation of the method in the second paper and start to write the abstract for SPWLA.
Summary:
I did literature review and did not find any paper using the same method as mine. I think it is innovative.
I start to write the abstract for SPWLA. I will try to finish it tomorrow and send it to you.
Tomorrow, I will continue to write the abstract for SPWLA.
Summary:
I did literature review and did not find any paper using the same method as mine. I think it is innovative.
I start to write the abstract for SPWLA. I will try to finish it tomorrow and send it to you.
Tomorrow, I will continue to write the abstract for SPWLA.
Subscribe to:
Posts (Atom)