i've been struggling what im going to start to make this function
im looking some plugins that can customize the Export Orders Data, and im using WooCommerce Customer/Order XML Export Suite Extension, And my Company they want to generate it in TXT file not XML file because they need to import the orders data from online woocommerce to local on FoxPro System to update the Stocks in their warehouse inventory.
the format of export should be like this
Example 1
Date Quantity Customer Name Cost Assigned Employee
07/12/17 5 John Doe 5000 MIRIAM
No. Location Product Name
000001 USA iPhone S9 Plus
Example 2
Varation Price Quantity
PC 110 10
PC 200 5
BX 500 3
Thank you.
I know nothing about WooCommerce, but a little web searching tells me that it can Export its data in either XML or CSV (one version of TXT) format.
See: DRAG & DROP WOOCOMMERCE XML & CSV EXPORTS
Perhaps the CSV exported data will be enough to meet your needs.
But if not, I don't know if you have the ability to modify your Foxpro/Visual Foxpro application. If so, you can easily create a utility within your FP/VFP application to either read & import the XML data or to import CSV data.
Once you better understand what options you have, let us know and we can most likely advise you better.
Good Luck
Related
I am fairly new to WooCommerce, so maybe I'm just missing a point. My client gave me an url to a csv file, that's updating every 15 Min.
The url is structured like: www.domain.com/backend/export/index/products-simple.csv?feedID=25&hash=... followed by a long letter/number combination
So my questions are:
How can I import the products? The link doesn't end with .csv, so the normal importer doesn't accept the url.
How can I update the stocks and products automatically every 15 Min?
For everyone interested in the solution: instead of creating a custom script, I used the WordPress all-import plugin: https://www.wpallimport.com/
There I could easily connect custom cron jobs and manage all the imported data.
I am trying to export payment details in a below report format:
Date Agst Ref Party Name Amount
11/25/2019 19-20/1256 ABC 4,145
Extracting Day Books / Transaction Data from Tally using ODBC
The answer in the above link solved 75% of my problem. I was able to extract Date, Party's Ledger Name and Amount. But I am struggling to export the 4th data point "Agst Ref" for both payment and receipt.
I am fine with sending XML export request if exporting it via ODBC route is not possible. In which case I need XML code.
Please assist me because I am stuck on this for a while.
Agst Ref is just $Reference. What do you get when you run the queries?
I have a requirement to load the csv into DB using oracle apex or pl/sql code, but the problem is they are asking to load the csv file which will not come with same number of columns and column names .
I should create table & upload data dynamically based on the file name and data that i'm uploading.
For every file i need to create a new table dynamically and insert data that are present in csv file.
For Example:
File1:
col1 col2 col3 col4 (NOTE: If i upload File 1, Table should be created dynamically based on the file name and table should contain same column name and data same as column headers of csv file . )
file 2:
col1 col2 col3 col4 col 5
file 3:
col4 col2 col1 col3
Depending on the columns and file name i need to create table for every file upload.
Can we load like this or not?
If yes, Please help me on this.
Regards,
Sachin.
((Where's the PL/SQL code in this solution!!??! Bear with me... the
answer is buried in here somewhere... I introduced some considerations
and assumptions you will need to think about before going into the
task. In the end, you'll find that Oracle APEX actually has a
built-in solution that satisfies exactly what you've specified... with
some caveats.))
If you are working within the Oracle APEX platform, you will have some advantages. APEX Version 4.2 and higher has a new page element called "Data Loading". The disadvantage however is that the definition of the upload target is fixed and not dynamic. You will need to know how your table is structured prior to loading the data.
One approach to overcome this is to build a generic, two-column table as your target, which will serve for all uploads. Column 1 will be your file-name and column two will be a single clob data type, which will contain the entire data file's contents including the header row. The "Data Loading" element will give the user the opportunity to verify and select this mapping convention in a couple of clicks.
At this point, it's mostly PL/SQL backend work doing the heavy lifting to parse and transform the data uploaded. As far as the dynamic table creation, I have noticed that the Oracle package, DBMS_SQL allows the execution of DDL SQL commands, which could be the route to making custom tables.
Alex Poole's comment is important as well, you will need to make some blanket assumption about the data type or have a provision to give more clues about what kind of data is contained. Assuming you can rely on a sample of existing data values is not good... what if all the values in your upload are null? I recommend perhaps a second column in the data input with a clue about the type of data for each column... just like the intended header names, maybe: AAAAA = for a five character column, # = for a numeric, MM/DD/YYYY = for a date with a specific masking.
The easier route:
You will need to allow your end-user access to a developer-role account on a workspace of your APEX server. It is not as scary as you think. With careful instruction and some simple precautions, I have been able to make this work with even the most non-technical of users. The reason for this is that there is a more powerful upload tool found under the following menu item:
SQL Workshop --> Utilities --> Data Workshop
There is a choice under "Data Load" --> "Spreadsheet Data"
The data load tool will automatically do the following:
Accept a CSV formatted file through a browse function on your client machine
Upload the file and parse the first record for the column layout (names)
Allow the user to create a new table from the uploaded file, or to map to an existing one.
For new tables, each column data type can be declared and also a specific numeric/date mask if additional conversion from the uploaded data is necessary.
Delimiter type, optional enclosures (like double quotes), decimal conventions and currency types can also be declared prior to parsing the uploaded file.
Once the user has identified all these mappings and settings, the table is created with the uploaded data. Any errors in record upload are reported immediately afterwards with detailed feedback on the failed records.
A security consideration to note:
You probably do not want to give end users access to your APEX server's backend... but you CAN create a new workspace... just for your end users... create a new database schema for receiving their uploads, maybe with some careful resource controls. Developer is the minimum role needed... but even if the end users see the other stuff there won't be access to anything important from an isolated workspace.
I have implemented the isolated workspace approach on a 4.0/4.1 release APEX platform a few years back, and it worked nicely. Our end user had control over the staging and quality checking of her data inputs (from excel spreadsheet/csv exports collected from a combination of sources). I suppose it may have been even better to cut her out of the picture entirely and focused on automating the export-review-upload process between our database and her other sources. In this case, the volume of data involved was not great enough (100's to 1000's of records) and the need for manual review and edit of the exported data was very important prior to pushing it into the database... so the human element was still important in this case - it is something you'll want to think about now.
What is the best practice to upload bunch of data(multiple rows) at once time. I don't want to upload any files on the server.
Is it good to have a text area to input data with a predefined structure(format). And create a small parser to read and analyze that input to insert it into to the database.
Edit:
I have the data set in excel file. I want to store it in the database, I don't want to upload the server.
Data sample :
id fid sid name
--------------------------------------------------------------
1- 3a3458 2a2125 3a4541 John Smith
2- 313547 3a4541 212145 Albert koku
.....................
...............
.........
100- ...
Since the data has already a predefined structure, to avoid manual errors and parsing I would build an interface which contains a <table> and the user will fill the corresponding data and then submit it to the server.
But if you don't want to bother to guide the user and help him then you could of course use a <textarea> in which the user could enter the data under some known form: CSV, JSON, XML, ... and then do the parsing on the server.
I'm trying to find the documentation for how the flat file looks for modifying the quantity of a product on Amazon.
This is what we send at the moment but it would be good to see what the list of headings we can use.
SKU | Quantity
000 | 1
I'm guessing that this is correct,
SKU | Price | Quantity
000 | 9.99 | 1
Any links would be welcome.
Amazon's MWS site https://developer.amazonservices.com/
You can get to the full description of the flat file feed specifications to send the flat file in the correct format to the MWS by going to the following URL after you're logged into your sellercentral or mws account.
https://sellercentral.amazon.com/gp/help/help.html/ref=ag_13461_cont_help?ie=UTF8&itemID=13461&language=en_US
have a look at the scratchpad, this will bring back flat files so you can see how they look.
https://mws.amazonservices.co.uk/scratchpad/index.html - UK Version
https://mws.amazonservices.com/scratchpad/index.html - General (US) version