I have a scenario in which I have to export entire information from the Alfresco and import that into another alfresco...
To not find any difference in users, sites, dashlets, rules, aspects and other information. It should work without any difference as like the previous one after import.
I used the plugin share-import-export-0.1-JAR-alfresco-3.4 for import and export in the Alfreso share itself. But my export option in share is not working consistently, sometimes it is successfully exported and some times showing error.
After export if I get the .acp file successfully... I am getting error msg while importing the entire content in share and the error msg also not so descriptive. Err Msg:"Unexpected error occured during content extraction".
You can't use an ACP for your needs. An ACP includes the content and permissions, but not the definitions of the users/groups of those permissions. An ACP also won't include the site definition - there's more to a site than just the content that makes it up.
If you need to export a site, including its contents and its users, then your best bet is to use the new functionality that was introduced into Alfresco 3.5 (Team) to support the sample site. There's a webscript that handles the export, and a bootstrap importer that'll load it into a new system for you. I find it works pretty well for this sort of thing, but then I'm possibly biased as I wrote most of it ;-)
The webscript for the export is org.alfresco.repository.site.site-export.get, which lives at /alfresco/services/api/sites/{shortname}/export - drop the site name into the URL and fetch (as an admin) to receive the zip with all the parts of the site in it.
To see the loading, take a look at patch-services-context.xml and search for patch.siteLoadPatch.swsdp. The class that loads is org.alfresco.repo.admin.patch.impl.SiteLoadPatch and it takes the users, people, groups, content and site name as parameters. (You'll get all of these things in the zip file from the export)
If you need to do multiple sites + users + groups + contents, then you should look at something like import/export, replication/transfer, or rolling something custom yourself (probably based on CMIS).
Related
I have an issue and I have been struggling with for weeks, almost a month, I am working on this website for a real estate agent in Toronto, Ontario and the last thing I have to do is get the listing on her website. We are using Wordpress wp-residence theme, this theme is compatible with iHomefinder, however the data I get from iHomefinder does not work with the themes features (maps, searches and even styles) I read somewhere that the only way to use these theme features is to add the listing manually. I found this plugin that will import all the listings
https://en-ca.wordpress.org/plugins/wp-residence-add-on-for-wp-all-import/
But now I need to get my listings in the form of XML, CSV or XLS. I have all my login info and url to get my listing, however in the instructions it says I need to connect via RETS Client, which I do not have...is there anyone out there who can point me in the right direction?
You'll need to use an IDX plugin.
I think you should also look over https://mlsimport.com .This plugin should help you import data from MLS. But is a paid version.
I have recently started developing with aikau in alfresco share.
I want to achieve a functionality wherein I can export search results to a CSV file.
For that, I can change the back-end repository web script to return csv data.
Now, At alfresco share end - I was successfully able to show the export link by adding a new widget to FCTSRCH_TOP_MENU_BAR. I used alfresco/renderers/PropertyLink to display this link. Now, the missing part for me is - how can I invoke the search web script passing additional param format=csv and alongwith that pass all the query parameters used to retrieve the results.
I am stuck with that. If I use the publishTopic as ALF_CRUD_GET_ALL and provide the URL there then it invokes the sample web script (I created to return sample csv response) and returns the response. However, the csv doesn't come as downloadable response. I am stuck here in order to how to achieve export csv functionality for search results.
It would be great if any of you can help me here and provide your guidance/suggestions.
This blog post provides an example on how you can custom the search page in Share. Although it specifically addresses changing the search queries the basic extension approach is more or less that same in that you will want to change the data that is used to send an XHR request. I think that the major difference here is that you may need to do more in-depth updates to the service - in particular with regards to the switch statement that is used to build the advanced search query object.
If you have extended or replaced the default search REST API then I would expect that you will need to call the same URL, but if you have provided an entirely new REST API to return the CSV data then you'll also need to change the URL that is used by the service.
In terms of providing a link for downloading the content we have previously implemented something in the DragAndDropModelCreationService (see the generateDownload function) but this only works with Chrome due to security limitations and the generation of files to download.
Your best bet may be temporarily store the CSV content on the repository in a hidden location and then use the standard download links to allow it to be downloaded - this would be more complex but would provide better cross browser support. Something similar is done for the "Download as ZIP" action.
OK, with the extra information provided I would do the following...
The information on the process of adding widgets to the search page are quite well detailed here (although you're not adding a view, you can follow the approach to add a new PropertyLink after the widget with the id "FCTSRCH_RESULTS_COUNT_LABEL").
The approach I would take would be to include an additional custom service on the page that subscribes to the "ALF_RETRIEVE_DOCUMENTS_REQUEST_SUCCESS" topic (which is published on a completed search). It should save the the search response in a variable in preparation for users clicking on the PropertyLink.
This custom service should also subscribe to a topic that is published by the PropertyLink (called say "DOWNLOAD_CSV"). This custom service could then generate a file download using the approach described in my previous answer using the CSV data that will have been provided in the payload. As I said though, this may only work with some browsers due to security reasons.
If your custom search WebScript were able to store the CSV data as a node on the Repository then you could just provide a the NodeRef of the CSV data in the search response and the PropertyLink could just publish the "ALF_DOWNLOAD" topic for the DocumentService to handle the download.
Trying to generate a file to download on the client side is going to be an issue for most browsers I think.
I'm running plone 3.3.5
I use Quills for blogging and need to export the entries to either XML of CSV.
I tried with the standard export to xml in zope, but this is not working.
So now I want to use products.csvreplicata, installation was succesfull, but I don't know how to use is.
ZMI export in XML is broken and it's however a very low level export, unreadable by human (do not use it).
CVSReplicata is an add-on targeted for developers, not for final users that want to export. It mean you must write some code (as the documentation says: see interfaces and adapters).
I am working on a directory site based on Wordpress and I need to import categories and members from an existing website. WordPress currently has an importer tool that is said to allow imports of categories...but it's in the WXR format.
From my understanding, a WXR is a XML file, but I'm not sure the exact format to use for the category and member imports (if that is even possible).
I searched different sources for a solution to this but couldn't find anything.
I have over 70 categories and over 600 business members to import. I would appreciate any advice on importing multiple categories & members into WordPress.
Thanks a lot
I found a solution to import multiple categories. I used the WordPress export and exported all the information from my site (posts, categories, ect). Then I carefully looked through the xml file and found the type of data I wanted to import (which was categories). So once I found the categories and its format, I removed everything else in between the ... tags, after the file details (titles, link, description, pubDate, generator, ect).
The category format is:
<wp:category><wp:category_nicename>"This is the Slug name" (no quotes</wp:category_nicename><wp:category_parent>"This is the parent name, optional" (no quotes)</wp:category_parent><wp:cat_name><![CDATA["This is the Category name" (no quotes)]]></wp:cat_name></wp:category>
You would create as many of this example as you need and just change the required information for the import. I think this strategy can also be used for other data to import. You would just have to examine the structure of the data in the original xml exported file and basically copy what you see.
Im still open to other techniques to import data. I hope this helps someone with importing needs.
I have requirement where I need to allow users to upload a Word document with place holders for certain fields which can be found in the database. This will be their template. For example the place holders might be prepended with ## or something. For example
Dear ##Title ##Lastname
They then can grab a record and hit export to Word document. This will then let them choose the template. They can select the template and then click continue. I will then get the template and replace the ##Title with the title field in the database for the selected record. I am not sure where to start or what components I need to do this.
From my initial investigation it seems that I can do this with the new open XML standard for Office 2007. So perhaps I should read in the template and save all the contents to a db table somewhere. Then when the use wants to export I get the contents again and then do a search and replace for the ## placeholders and link them properly. Then save the document to the output stream again which will then bring up the save dialog on their browser.
I am using ASP.Net MVC and am in a hosted environment. I was also maybe contemplating dynamically creating a new View type and dynamically creating new views when the user uploads a template. Not sure that this approach will work though.
Is this a good approach?
What tools should I be looking at?
Any other suggestions?
This is similar to an approach we took for inserting data into word documents and then returning them to the user. We opened the .docx file (it's a zip file so easy to extract) extracted the document (in the word folder called document.xml), did the replace and then put the document back into the .docx file and returned it to the user.
An issue we hit were that word inserted tags at strange places, especially things like spell/grammar errors, so we needed to be careful when we did the search/replace.
We decided not to store the fields from the document in a database to allow the documents to be easily updated.
We used dotnetzip component for opening the .docx files
Something we also did was to combine several documents into a single large document to save on the number of downloads. If I remember we used the open xml toolkit to do this merging. The website has also got loads of other information that may be of use.
Check out Scott Guthries blog post about the new view engine code named "Razor" coming out real shortly from Microsoft. In the comments there is talk about it being able to be used in mail merge scenarios like you talk about with ASP.NET MVC views.