I have a .csv that was created using input_xls.to_csv, and everything delimits correctly. However, within some of the string records are commas (e.g. a business name like 'Widgets, Inc.'). The result is incorrect delimiting when importing the .csv into a file geodatabase using arcpy.conversion.TableToTable - values end up shifted over into other columns. I want to replace the commas with a space.
I cannot figure out how to use either something like
.apply(lambda x: x.replace(',', ' '))
[val if val else " " for val in row] + ([" "] * (len(first_row) - len(row))) (found here: https://community.esri.com/t5/python-questions/find-and-replace-blanks-in-csv-with-python/m-p/673632)
Any help to nudge me in the right direction would be appreciated. Thanks.
What you can do is read your excel file into an array, where when appending data into the array, you call the replace command to get rid of the comma.
inRow = [str(row),str(row),str(row),str(row),str(row),str(row),str(row),str(row),str(row),str(row),str(row)]index = 0for i in inRow:if "," in i:i = i.replace(",","")inRow[index] = iindex += 1points.append(inRow)
AllProjects_CSV.truncate()writer = csv.writer(AllProjects_CSV,delimiter=',')writer.writerow(fields)writer.writerows(points)
You might want to try bringing in the csv back into Excel, then try the Excel to Table tool to see if it handles the embedded commas better when bringing back into a file geodatabase
Personally I cannot recreate your probleme, but you have two options.
1. change regional settings of your machine to use different seperators (semicolon for example)
2. in the same folder as your data add file "schema.ini", where you specify how data is seperated, additionnaly you can change the name of columns and its types.
I never followed-up with the solution I found using csv and pandas:
Creating a new .csv is not an issue, because I delete all of them at the end of the process.