I went to old espo (5.8) and exported all entities of a custom scope with all the fields. Then I started a cron import in the new espo instance (5.9) but it hangs after 654 entities out of 25000 and stuck in "in progress". I was thinking if the file was created by espo it would not have problems importing. This is a bug I think....
How to troubleshoot import of CSV files created by Espo ?
Collapse
X
-
-
I had multiple problems (see the email thread) that spammed the log file so it is really big now (18GB). I am trying to trim duplicate lines in order to make it more manageable size and see if something comes up.
Thinking about this, the problem may be in either csv parser or export code... but we need details.
The data looks quite "normal". All these entries have multiline descriptions (the third column is a normal "description" field) but the previous ones were imported without problems...
I will create a csv with only that portion so I can try to reproduce the problem.
I was noticing the ones in the third column are CRLF while the ending lines are LF. Also there are TAB characters in there.
It may have to do with that but why would the first batch of entities be imported without issue?
Ps. I was thinking... wouldn't encoding the values with hex/url/base64 solve most problems of csv files?
In that case when importing or exporting we could have a checkbox to do it or decode when importing. maybe I will try to implement it as a custom format.Last edited by tothewine; 05-22-2020, 11:58 AM.Comment
-
It is very strange... the second CSV file imported without errors.... while the first is still "in progress".
When doing the import without idle mode I saw a gateway error. The idle mode should not have this problem.
The bug here is that I see no way to resume an import without making a new one...
Last edited by tothewine; 05-24-2020, 07:07 PM.Comment
-
After much digging I understood the php timeout limits were different between cron jobs and apache. This is now a bug report about having some logic for tracking csv import and resume it automatically. for example storing the import settings in the import entity rather than the import job and then having an 'offset' field in there which is increase by 1 every time an entity is imported correctly. If the import is interrupted then it would be sufficient to hit a "resume" button that will restart the import job, skipping the number of entities in 'offset' before start the actual import.Last edited by tothewine; 05-25-2020, 12:45 PM.Comment
-
Comment