I have installed alfresco community 4.2 on my laptop. Also i have downloaded alfresco workdesk & configured it to use my local alfresco repository instead of default one.
Now i am not able to figure out how to get workflow features on content i uploaded in workdesk UI as well as how to proceed with case management features as clubbing multiple files to create one case etc.?
Can anybody help me here?
OK. I have done some more research & found out that one can use alfresco enterprise edition & install workdesk enterprise trial for same. This trial is a preconfigured case management example which has lots of features which can be used as yardsticks such as creating case folders with multiple files, workflow etc.
Now depending upon information/documentation you have got, one can extend it for further use.
Related
We are upgrading from alfresco 5.0.1 to alfresco 5.2.2.
Actually it's not just upgradation, but it's a consolidation of 3 different projects.
Currently we have 3 projects in alfresco. 1st one is in alfresco 5.0.1, 2nd one is in alfresco 5.2 and 3rd one is in 5.0.
And here we have 3 separate repositories.
Now we are consolidating these 3 projects.
So we are going to keep only one repo. for all three projects and there will be 3 separate share for each application.
And in this case will keep one repository as is, but other 2 repositories we need to migrate.
So in above scenario after up-gradation we want to migrate actual repository contents from old alfresco repository to new alfresco repository with all associated metadata.
So please let me know if alfresco is having any in-build feature to achieve this requirements easily.
Also let me know, what are the other possible ways to achieve this.
All your inputs will be really appreciated.
Thanks in Advance.
There is nothing out-of-the-box that will help you consolidate three separate repositories into a single repository.
Your best bet is likely to write an export script that will write the files out to the file system along with an XML file that describes the metadata. It will simplify things if you write the XML file in the same format that the Bulk File System Import Tool (BFSIT) expects when importing.
Write the data to a volume that can be mounted by the Alfresco server you are consolidating into.
Once the data is written and the volume is mounted, you can run a BFSIT import to bring the files and the metadata into the repository.
Using this approach you probably can avoid upgrading the 5.0.x servers.
This only addresses data. Of course you'll also need to deal with customizations you've made to the two 5.0.x servers. Hopefully, everything you've done in those other two servers are namespaced and isolated so that they won't collide with the third server. You'll need to test your 5.0.x customizations to make sure they will work with 5.2.x before cutting over.
In Biztalk, many artifacts (e.g. schema, pipeline etc) are used in some location like sendport, receivelocation etc.
Each time when we want to deploy an update version to Biztalk, it will said the artifact is being used. But we dont know where it is being used and we need to check all receivelocations, sendports and other places one by one.
Is there any faster approach such that we can know where the artifact is used?
Thanks.
BizTalk documenter might be of some use. It documents in what ports a map is used. (Full Disclosure: I am now a developer on the latest versions)
And just released there is a new tool called BTSG NoS Addin will find your external dependencies.
You can check this using Admin Console of Biztalk.
For example if you want to know where a pipeline is used, just right click on pipeline and click on 'view dependencies' and now you can view all those places where pipeline is being used/refered. click on Number and it will bring to that place
I'm very new to Alfresco and trying to get hold of it. I'm using Community edition 4.2 and I've created one website using Alfresco Share and I did not imported Government/Finance website data. Below is the screenshot -
Now I've created two folders under Document Library tab and have uploaded few PDFs into that. It looks like below -
Now if I click on publish link then its saying that "There is no appropriate channel for this content".
I entered ubswcm as URL name while creating this site.
I followed the link http://docs.huihoo.com/alfresco/Getting-Started-with-WCM-for-Alfresco-Commumity-Edition-3.2.pdf which talks about stating but I could not find staging in this version of Alfresco 4.2.
My questions
How can I publish this content?
How can I run this website on localhost? Do I need to create war file? If so then please advice how to proceed on this.
http://:8080/wcmqs running fine. How can I build slimier site like this in http://:8080/ubswcm ? Please suggest if you have any tutorial on this.
Many thanks for your help!
You need to follow a tutorial on the Alfresco Web Quick Start capabilities. You should find that these capabilities are well-documented on docs.alfresco.com, e.g. Using Alfresco Web Quick Start.
Do not confuse Web Quick Start with the Social Publishing framework (which the Publish action is associated with) or with the legacy AVM-based WCM services (which I assume the PDF refers to, but I could not tell for sure as you link gives me a 404) that are no longer supported in Alfresco 4.
When I create a new Drupal site I usually end up with at least one custom module and several community contributed modules. To get the site working as it should, many configuration values need to be set on the various modules. This makes deployment onto a fresh Drupal instance painstaking and error-prone.
I would like to give my custom module the ability to configure all the other modules. Either on install or on the click of a button on my custom module's administration page, all the necessary configuration values on the other modules would be programmatically set.
How would I best go about doing this?
AFAIK, there's no way to achieve what you mean easily. I tend to put as much as I can in hook_update_N() implementations and do frequent DB synchronisations as described in my answer to this question. However that does not work when you already have a live server with which you will have to merge data.
To that purpose, I use various tools according to the need. No one is perfect, but here's however a small collection of my favorite ones:
Features. This is a new concept and a new module. The idea is pretty awesome: it allows you to define a set of configuration/modules/settings and to export them as a feature. This feature will then be installed as if it were a module on the target site. This module does not export every possible setting, but it does however do a good job with the modules that need the hardest configuration, as CCK, Views, ImageCache and others... You can see a screencast demo (~10 mins) here.
Backup and migrate. This is a more radical approach: it simply dump and rebuild the entire database on a target system. It is good only if you need to overwrite the target system completely.
Node export. This allows to export (and import) nodes from a drupal installation to another one. It supports bulk operations but - unluckily - it does not support the migration of attached files and images.
Deploy. Because of the limitations of node export I once looked into using this module (still in development). I finally did not, and preferred to do a merge of the production and staging databases, but the concept seems very valid, as it allows to import/export complex data type via SOAP.
Taxonomy import/export. I suppose the name is self-explanatory. It uses files to achieve the tasks (XML or CSV).
Installation profiles (suggested by ctford) are useful when configuring new sites. They allow you to specify modules to enable, theme to default to etc on installation. They can be quite convenient because there is a command-line tool called Drush that automates the building of installation profiles. The downside is that the profiles are designed to be used on installation - not deployment of an individual module. It might be possible however to take the configuration code generated by Drush and call it when your module is enabled.
Finally, you can find a collection of tools for importing/exporting data here.
HTH!
have you looked at the "features" module? it is a new paradigm introduced as part of the open atrium distribution but also available as a stand-alone module. from their description:
"The features module enables the capture and management of features in Drupal. A feature is a collection of Drupal entities which taken together satisfy a certain use-case.
Features provides a UI and API for taking different site building components from modules with exportables and bundling them together in a single feature module. A feature module is like any other Drupal module except that it declares its components (e.g. views, contexts, CCK fields, etc.) in its .info file so that it can be checked, updated, or reverted programmatically."
http://drupal.org/project/features
Installation profiles are useful when configuring new sites. They allow you to specify modules to enable, theme to default to etc on installation. They can be quite convenient because there is a command-line tool called Drush that automates the building of installation profiles.
The downside is that the profiles are designed to be used on installation - not deployment of an individual module. It might be possible however to take the configuration code generated by Drush and call it when your module is enabled.
I know what you mean, it's a pain to set all modules up.
I'm sure you can investigate all 3rd party modules to see how configuration takes place and mimic that in your custom module, but I'd advise you against that...
The problem is that modules may change the way they store their settings from one revision to another, so whenever you update to a new version of any module you should do some reverse-engineering to see if your 'ultimate-one-click-configuration module' still works ok - which, if you ask me, is even more painful than manually configuring all modules for each project.
Just relax, take it easy, and enjoy Drupal :)
As the initialization is only required when Drupal is installed, I would think that a installation profile is the better solution; to keep a module that is not anymore used once that the installation is configured seems a little excessive, IMO.
Changing the installation profile used from a site, and make the new installation profile run its installation code isn't something that Drupal allows out-of-the-box. I would create a custom installation profile before creating the sites I need, and only for the features I know all the sites will share. For the other features, I would create separate custom modules I can later install, and eventually uninstall when the features they implement aren't anymore necessary.
What are the strategies for versioning of a web application/ website?
I notice that here in the Beta there is an svn revision number in the footer and that's ideal for an application that uses svn over one repository. But what if you use externals or a different source control application that versions separate files?
It seems easy for a Desktop app, but I can't seem to find a suitable way of versioning for an asp.net web application.
NB I'm not sure that I have been totally clear with my question.
What I want to know is how to build and auto increment a version number for an asp.net application.
I'm not interested in how to link it with svn.
I think what you are looking for is something like this: How to auto-increment assembly version using a custom MSBuild task. It's a little old but I think it will work.
For my big apps I just use a incrementing version number id (1.0, 1.1, ...) that i store in a comment of the main file (usually index.php).
For just websites I usually just have a revision number (1,2,3,...).
I have a tendency to stick with basic integers at first (1,2,3), moving onto rational numbers (2.1, 3.13) when things get bigger...
Tried using fruit at one point, that works well for a small office. Oh, the 'banana' release? looks over in the corner "yeah... that's getting pretty old now..."
Unfortunately, confusion started to set in when the development team grew, is it an Orange, or Mandarin, or Tangelo? It looks ok. What do you mean "rotten on the inside?"
... but in all honesty. Setup a separate repository as a master, development goes on in various repositories. For every scheduled release everything is checked into the master repository so that you can quickly roll back when something goes wrong.
(I'm assuming dev/test/production are all separate servers, and dev is never allowed to touch production or the master repository....)
I maintain a system of web applications with various components that live in separate SVN repos. To be able to version track the system as a whole, I have another SVN repo which contains all other repos as external references. It also contains install / setup script(s) to deploy the whole thing. With that setup, the SVN revision number of the "metarepository" could possibly be used for versioning the complete system.
In another case, I include the SVN revision via SVN keywords in a class file that serves no other purpose (to avoid the risk of keyword substitution breaking my code). The class in that file contains a string variable that is manipulated by SVN and parsed by a class method.
An inconvenience with both approaches is that the revision number is not automatically updated by changes in the externals (approach 1) or the rest of the code (approach 2).
During internal development, I'm using milestone numbers (M1, M2, M3...). After release, I'll probably just update dates ("the January 2009 update").