HTTP Status 404 on /alfresco - alfresco

I have installed Alfresco on port 8180 and it's not running. It's giving error HTTP status 404 at /alfresco.
alfresco.log output is:
13:56:20,665 INFO [org.alfresco.repo.admin] Using database URL 'jdbc:mysql://localhost:3306/alfrescobo' with user 'alfrescobo'.
13:56:20,899 INFO [org.alfresco.repo.admin] Connected to database MySQL version 5.5.47-0ubuntu0.14.04.1
13:56:27,524 INFO [org.alfresco.repo.management.subsystems.ChildApplicationContextFactory] Starting 'sysAdmin' subsystem, ID: [sysAdmin, default]
13:56:27,549 INFO [org.alfresco.repo.management.subsystems.ChildApplicationContextFactory] Startup of 'sysAdmin' subsystem, ID: [sysAdmin, default] complete
13:56:36,184 INFO [org.springframework.extensions.webscripts.TemplateProcessorRegistry] Registered template processor Repository Template Processor for extension ftl
13:56:36,186 INFO [org.springframework.extensions.webscripts.ScriptProcessorRegistry] Registered script processor Repository Script Processor for extension js
13:56:42,040 INFO [org.alfresco.repo.domain.schema.SchemaBootstrap] Connecting to database: jdbc:mysql://localhost:3306/alfrescobo, UserName=alfrescobo#localhost, MySQL Connector Java
13:56:42,041 INFO [org.alfresco.repo.domain.schema.SchemaBootstrap] Schema managed by database dialect org.hibernate.dialect.MySQLInnoDBDialect.
13:56:42,729 INFO [org.alfresco.repo.domain.schema.SchemaBootstrap] No changes were made to the schema.
13:56:43,239 ERROR [org.alfresco.repo.dictionary.DictionaryRepositoryBootstrap] Failed to create model (due to EmptyContentReader): workspace://SpacesStore/154d07f2-fc85-4625-8197-6bfb7348a0ec
13:56:43,273 INFO [org.alfresco.repo.management.subsystems.ChildApplicationContextFactory] Starting 'Authentication' subsystem, ID: [Authentication, managed, alfrescoNtlm1]
13:56:43,417 INFO [org.alfresco.repo.management.subsystems.ChildApplicationContextFactory] Startup of 'Authentication' subsystem, ID: [Authentication, managed, alfrescoNtlm1] complete
13:56:43,684 INFO [org.alfresco.repo.management.subsystems.ChildApplicationContextFactory] Starting 'Search' subsystem, ID: [Search, managed, solr]
13:56:43,941 INFO [org.alfresco.repo.management.subsystems.ChildApplicationContextFactory] Startup of 'Search' subsystem, ID: [Search, managed, solr] complete
13:57:19,184 ERROR [org.springframework.web.context.ContextLoader] Context initialization failed
org.alfresco.error.AlfrescoRuntimeException: 02040119 Workflow deployment failed
at org.alfresco.repo.workflow.WorkflowDeployer.init(WorkflowDeployer.java:345)
at org.alfresco.repo.workflow.WorkflowDeployer$1$1.doWork(WorkflowDeployer.java:469)
at org.alfresco.repo.security.authentication.AuthenticationUtil.runAs(AuthenticationUtil.java:548)
at org.alfresco.repo.workflow.WorkflowDeployer$1.execute(WorkflowDeployer.java:465)
at org.alfresco.repo.transaction.RetryingTransactionHelper.doInTransaction(RetryingTransactionHelper.java:452)
at org.alfresco.repo.workflow.WorkflowDeployer.onBootstrap(WorkflowDeployer.java:460)
at org.springframework.extensions.surf.util.AbstractLifecycleBean.onApplicationEvent(AbstractLifecycleBean.java:56)
at org.alfresco.repo.management.SafeApplicationEventMulticaster.multicastEventInternal(SafeApplicationEventMulticaster.java:209)
at org.alfresco.repo.management.SafeApplicationEventMulticaster.multicastEvent(SafeApplicationEventMulticaster.java:180)
at org.springframework.context.support.AbstractApplicationContext.publishEvent(AbstractApplicationContext.java:303)
at org.springframework.context.support.AbstractApplicationContext.finishRefresh(AbstractApplicationContext.java:911)
at org.springframework.context.support.AbstractApplicationContext.refresh(AbstractApplicationContext.java:428)
at org.springframework.web.context.ContextLoader.createWebApplicationContext(ContextLoader.java:276)
at org.springframework.web.context.ContextLoader.initWebApplicationContext(ContextLoader.java:197)
at org.springframework.web.context.ContextLoaderListener.contextInitialized(ContextLoaderListener.java:47)
at org.alfresco.web.app.ContextLoaderListener.contextInitialized(ContextLoaderListener.java:63)
at org.apache.catalina.core.StandardContext.listenerStart(StandardContext.java:4939)
at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5434)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:150)
at org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:901)
at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:877)
at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:633)
at org.apache.catalina.startup.HostConfig.deployWAR(HostConfig.java:976)
at org.apache.catalina.startup.HostConfig$DeployWar.run(HostConfig.java:1653)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
at java.util.concurrent.FutureTask.run(FutureTask.java:166)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:724)
Caused by: org.alfresco.service.cmr.repository.ContentIOException: 02040118 Failed to open stream onto channel:
accessor: ContentAccessor[ contentUrl=store://2016/3/2/7/32/78b6eff0-8082-4ed5-9d43-d3adad1f6a6b.bin, mimetype=text/xml, size=0, encoding=UTF-8, locale=en_US]
at org.alfresco.repo.content.AbstractContentReader.getContentInputStream(AbstractContentReader.java:401)
at org.alfresco.repo.workflow.WorkflowServiceImpl.isDefinitionDeployed(WorkflowServiceImpl.java:273)
at sun.reflect.GeneratedMethodAccessor341.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.springframework.aop.support.AopUtils.invokeJoinpointUsingReflection(AopUtils.java:309)
at org.springframework.aop.framework.ReflectiveMethodInvocation.invokeJoinpoint(ReflectiveMethodInvocation.java:183)
at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:150)
at org.alfresco.service.cmr.workflow.WorkflowPermissionInterceptor.invoke(WorkflowPermissionInterceptor.java:54)
at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:172)
at org.alfresco.repo.security.permissions.impl.ExceptionTranslatorMethodInterceptor.invoke(ExceptionTranslatorMethodInterceptor.java:46)
at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:172)
at org.alfresco.repo.audit.AuditMethodInterceptor.invoke(AuditMethodInterceptor.java:161)
at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:172)
at org.springframework.transaction.interceptor.TransactionInterceptor.invoke(TransactionInterceptor.java:110)
at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:172)
at org.springframework.aop.framework.JdkDynamicAopProxy.invoke(JdkDynamicAopProxy.java:202)
at com.sun.proxy.$Proxy57.isDefinitionDeployed(Unknown Source)
at org.alfresco.repo.workflow.WorkflowDeployer.deploy(WorkflowDeployer.java:368)
at org.alfresco.repo.workflow.WorkflowDeployer.init(WorkflowDeployer.java:324)
... 29 more
Caused by: org.springframework.dao.ConcurrencyFailureException: store://2016/3/2/7/32/78b6eff0-8082-4ed5-9d43-d3adad1f6a6b.bin no longer exists
at org.alfresco.repo.content.EmptyContentReader.getDirectReadableChannel(EmptyContentReader.java:56)
at org.alfresco.repo.content.AbstractContentReader.getReadableChannel(AbstractContentReader.java:271)
at org.alfresco.repo.content.AbstractContentReader.getContentInputStream(AbstractContentReader.java:375)
... 48 more

The file
alf_data/contentstore/2016/3/2/7/32/78b6eff0-8082-4ed5-9d43-d3adad1f6a6b.bin
has been deleted. We have no way to know what this file contained two day ago.
You will have to make a clean Alfresco installation and start again.

Make sure you are completely wiping out the database and the content store between each installation. If this log is from a fresh install, I think it may be missing the schema creation step, which means the database already exists which should not be the case for a fresh install.
If you are clearing out the content store but not the database, Alfresco will look for nodes that exist in the DB but no longer exist on the file system, which is essentially a corrupted repository.

Related

API Manager 4.1, Management Console cant force Restart

i have problem with cant force restart from management console login as admin. can anyone helping with me about this. i got log like this
`[2022-10-01 00:45:04,581] INFO - CarbonCoreActivator Starting WSO2 Carbon...
[2022-10-01 00:45:04,581] INFO - CarbonCoreActivator Operating System : Linux 4.18.0-305.el8.x86_64, amd64
[2022-10-01 00:45:04,581] INFO - CarbonCoreActivator Java Home : /usr/java/jdk1.8.0_341-amd64/jre
[2022-10-01 00:45:04,581] INFO - CarbonCoreActivator Java Version : 1.8.0_341
[2022-10-01 00:45:04,581] INFO - CarbonCoreActivator Java VM : Java HotSpot(TM) 64-Bit Server VM 25.341-b10,Oracle Corporation
[2022-10-01 00:45:04,582] INFO - CarbonCoreActivator Carbon Home : /opt/source/wso2_binary/wso2am-4.1.0
[2022-10-01 00:45:04,582] INFO - CarbonCoreActivator Java Temp Dir : /opt/source/wso2_binary/wso2am-4.1.0/tmp
[2022-10-01 00:45:04,582] INFO - CarbonCoreActivator User : abhimata, en-US, Asia/Jakarta
[2022-10-01 00:45:04,786] INFO - DefaultCryptoProviderComponent 'CryptoService.Secret' property has not been set. 'org.wso2.carbon.crypto.provider.SymmetricKeyInternalCryptoProvider' won't be registered as an internal crypto provider. Please set the secret if the provider needs to be registered.
[2022-10-01 00:45:05,125] INFO - KafkaEventAdapterServiceDS Successfully deployed the Kafka output event adaptor service
[2022-10-01 00:45:05,279] INFO - TemplateDeployerServiceTrackerDS Successfully deployed the execution manager tracker service
[2022-10-01 00:45:06,716] INFO - ServiceComponent Eventing Hub ServiceComponent is activated
[2022-10-01 00:45:07,431] WARN - Digester Match [Server/Service/Engine/Host/Valve] failed to set property [maxDays] to []
[2022-10-01 00:45:08,095] ERROR - DefaultRealm nullType class java.lang.reflect.InvocationTargetException
org.wso2.carbon.user.core.UserStoreException: nullType class java.lang.reflect.InvocationTargetException
at org.wso2.carbon.user.core.common.DefaultRealm.createObjectWithOptions(DefaultRealm.java:404) ~[org.wso2.carbon.user.core_4.6.3.jar:?]
at org.wso2.carbon.user.core.common.DefaultRealm.initializeObjects(DefaultRealm.java:231) ~[org.wso2.carbon.user.core_4.6.3.jar:?]
at org.wso2.carbon.user.core.common.DefaultRealm.init(DefaultRealm.java:136) ~[org.wso2.carbon.user.core_4.6.3.jar:?]
at org.wso2.carbon.user.core.common.DefaultRealmService.initializeRealm(DefaultRealmService.java:276) ~[org.wso2.carbon.user.core_4.6.3.jar:?]
at org.wso2.carbon.user.core.common.DefaultRealmService.<init>(DefaultRealmService.java:102) ~[org.wso2.carbon.user.core_4.6.3.jar:?]
at org.wso2.carbon.user.core.common.DefaultRealmService.<init>(DefaultRealmService.java:115) ~[org.wso2.carbon.user.core_4.6.3.jar:?]
at org.wso2.carbon.user.core.internal.Activator.startDeploy(Activator.java:72) ~[?:?]
at org.wso2.carbon.user.core.internal.BundleCheckActivator.start(BundleCheckActivator.java:61) ~[?:?]
at org.eclipse.osgi.internal.framework.BundleContextImpl$3.run(BundleContextImpl.java:842) ~[org.eclipse.osgi_3.14.0.v20190517-1309.jar:?]
at org.eclipse.osgi.internal.framework.BundleContextImpl$3.run(BundleContextImpl.java:1) ~[org.eclipse.osgi_3.14.0.v20190517-1309.jar:?]
at java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_341]
at org.eclipse.osgi.internal.framework.BundleContextImpl.startActivator(BundleContextImpl.java:834) ~[org.eclipse.osgi_3.14.0.v20190517-1309.jar:?]
at org.eclipse.osgi.internal.framework.BundleContextImpl.start(BundleContextImpl.java:791) ~[org.eclipse.osgi_3.14.0.v20190517-1309.jar:?]
at org.eclipse.osgi.internal.framework.EquinoxBundle.startWorker0(EquinoxBundle.java:1013) ~[org.eclipse.osgi_3.14.0.v20190517-1309.jar:?]
at org.eclipse.osgi.internal.framework.EquinoxBundle$EquinoxModule.startWorker(EquinoxBundle.java:365) ~[org.eclipse.osgi_3.14.0.v20190517-1309.jar:?]
at org.eclipse.osgi.container.Module.doStart(Module.java:598) ~[org.eclipse.osgi_3.14.0.v20190517-1309.jar:?]
at org.eclipse.osgi.container.Module.start(Module.java:462) ~[org.eclipse.osgi_3.14.0.v20190517-1309.jar:?]
at org.eclipse.osgi.container.ModuleContainer$ContainerStartLevel$1.run(ModuleContainer.java:1820) ~[org.eclipse.osgi_3.14.0.v20190517-1309.jar:?]
at org.eclipse.osgi.internal.framework.EquinoxContainerAdaptor$2$1.execute(EquinoxContainerAdaptor.java:150) ~[org.eclipse.osgi_3.14.0.v20190517-1309.jar:?]
at org.eclipse.osgi.container.ModuleContainer$ContainerStartLevel.incStartLevel(ModuleContainer.java:1813) ~[org.eclipse.osgi_3.14.0.v20190517-1309.jar:?]
at org.eclipse.osgi.container.ModuleContainer$ContainerStartLevel.incStartLevel(ModuleContainer.java:1770) ~[org.eclipse.osgi_3.14.0.v20190517-1309.jar:?]
at org.eclipse.osgi.container.ModuleContainer$ContainerStartLevel.doContainerStartLevel(ModuleContainer.java:1735) ~[org.eclipse.osgi_3.14.0.v20190517-1309.jar:?]
at org.eclipse.osgi.container.ModuleContainer$ContainerStartLevel.dispatchEvent(ModuleContainer.java:1661) ~[org.eclipse.osgi_3.14.0.v20190517-1309.jar:?]
at org.eclipse.osgi.container.ModuleContainer$ContainerStartLevel.dispatchEvent(ModuleContainer.java:1) ~[org.eclipse.osgi_3.14.0.v20190517-1309.jar:?]
at org.eclipse.osgi.framework.eventmgr.EventManager.dispatchEvent(EventManager.java:234) ~[org.eclipse.osgi_3.14.0.v20190517-1309.jar:?]
at org.eclipse.osgi.framework.eventmgr.EventManager$EventThread.run(EventManager.java:345) ~[org.eclipse.osgi_3.14.0.v20190517-1309.jar:?]
Caused by: java.lang.reflect.InvocationTargetException
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[?:1.8.0_341]
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) ~[?:1.8.0_341]
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[?:1.8.0_341]
at java.lang.reflect.Constructor.newInstance(Constructor.java:423) ~[?:1.8.0_341]
at org.wso2.carbon.user.core.common.DefaultRealm.createObjectWithOptions(DefaultRealm.java:358) ~[org.wso2.carbon.user.core_4.6.3.jar:?]
... 25 more
Caused by: org.wso2.carbon.user.core.UserStoreException: DB error occurred while persisting domain : PRIMARY & tenant id : -1234
at org.wso2.carbon.user.core.util.UserCoreUtil.persistDomain(UserCoreUtil.java:931) ~[org.wso2.carbon.user.core_4.6.3.jar:?]
at org.wso2.carbon.user.core.common.AbstractUserStoreManager.persistDomain(AbstractUserStoreManager.java:9083) ~[org.wso2.carbon.user.core_4.6.3.jar:?]
at org.wso2.carbon.user.core.jdbc.JDBCUserStoreManager.<init>(JDBCUserStoreManager.java:320) ~[org.wso2.carbon.user.core_4.6.3.jar:?]
at org.wso2.carbon.user.core.jdbc.JDBCUserStoreManager.<init>(JDBCUserStoreManager.java:262) ~[org.wso2.carbon.user.core_4.6.3.jar:?]
at org.wso2.carbon.user.core.jdbc.UniqueIDJDBCUserStoreManager.<init>(UniqueIDJDBCUserStoreManager.java:129) ~[org.wso2.carbon.user.core_4.6.3.jar:?]
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[?:1.8.0_341]
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) ~[?:1.8.0_341]
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[?:1.8.0_341]
at java.lang.reflect.Constructor.newInstance(Constructor.java:423) ~[?:1.8.0_341]
at org.wso2.carbon.user.core.common.DefaultRealm.createObjectWithOptions(DefaultRealm.java:358) ~[org.wso2.carbon.user.core_4.6.3.jar:?]
... 25 more
i expecting can force restart from management console or can with restart with command line from linux. please help me about this; thanks
Usually you can stop wso2server gracefully with ./wso2server.sh --stop. (./api-manager.sh --stop for latest versions)
In the management console(https://localhost:9443/carbon) also there is a shutdown button that you can use.
If you need to kill the process as you require right now, you can use in inbuilt Linux commands such as ps to locate the process and kill it.
Enter the command ps ax | grep wso2
This will list the set of processes started that has the text wso2. Unless you opened several apim instances, there will be one process that contains an ID. There, the first number refers to the process id.
Enter following command to gracefully shutdown the APIM.
kill <process-id>
You may force the shutdown by sending signal kill (-9) like follows if you need to.
kill -9 <process-id>
With ps again, if you can see there is a java process that sets the java home to <APIM_HOME>, additionally, you can kill it with
kill -9 java

Corda Settler on Corda4

I upgraded the Corda-Settler to the Corda 4 release and added the ripple libraries ripple-core and ripple-bouncycastle in order to make it independent on the local installation.
When I deploy the nodes, I get the following error:
[ERROR] 2019-02-12T11:02:31,343Z [main] internal.Node.installCordaServices - Corda service com.r3.corda.finance.obligation.oracle.services.XrpOracleService failed to instantiate. Reason was: No configuration setting found for key 'nodes' [errorCode=n3sk62, moreInformationAt=https://errors.corda.net/OS/4.0-RC03/n3sk62]
com.typesafe.config.ConfigException$Missing: No configuration setting found for key 'nodes'
at com.typesafe.config.impl.SimpleConfig.findKeyOrNull(SimpleConfig.java:152) ~[corda.jar:?]
at com.typesafe.config.impl.SimpleConfig.findOrNull(SimpleConfig.java:170) ~[corda.jar:?]
at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:184) ~[corda.jar:?]
at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:189) ~[corda.jar:?]
at com.typesafe.config.impl.SimpleConfig.getList(SimpleConfig.java:258) ~[corda.jar:?]
at com.typesafe.config.impl.SimpleConfig.getHomogeneousUnwrappedList(SimpleConfig.java:329) ~[corda.jar:?]
at com.typesafe.config.impl.SimpleConfig.getStringList(SimpleConfig.java:387) ~[corda.jar:?]
at com.r3.corda.finance.obligation.oracle.services.XrpOracleService$nodes$2.invoke(XrpOracleService.kt:22) ~[?:?]
at com.r3.corda.finance.obligation.oracle.services.XrpOracleService$nodes$2.invoke(XrpOracleService.kt:19) ~[?:?]
at kotlin.SynchronizedLazyImpl.getValue(LazyJVM.kt:74) ~[kotlin-stdlib-1.2.71.jar:1.2.71-release-64 (1.2.71)]
at com.r3.corda.finance.obligation.oracle.services.XrpOracleService.getNodes(XrpOracleService.kt) ~[?:?]
at com.r3.corda.finance.obligation.oracle.services.XrpOracleService.<init>(XrpOracleService.kt:24) ~[?:?]
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[?:1.8.0_191]
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) ~[?:1.8.0_191]
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[?:1.8.0_191]
at java.lang.reflect.Constructor.newInstance(Constructor.java:423) ~[?:1.8.0_191]
at net.corda.node.internal.AbstractNode.installCordaService(AbstractNode.kt:649) ~[corda-node-4.0-RC03.jar:?]
at net.corda.node.internal.AbstractNode.installCordaServices(AbstractNode.kt:579) ~[corda-node-4.0-RC03.jar:?]
at net.corda.node.internal.AbstractNode.access$installCordaServices(AbstractNode.kt:118) ~[corda-node-4.0-RC03.jar:?]
at net.corda.node.internal.AbstractNode$start$7.invoke(AbstractNode.kt:380) ~[corda-node-4.0-RC03.jar:?]
at net.corda.node.internal.AbstractNode$start$7.invoke(AbstractNode.kt:118) ~[corda-node-4.0-RC03.jar:?]
at net.corda.nodeapi.internal.persistence.CordaPersistence.inTopLevelTransaction(CordaPersistence.kt:227) ~[corda-node-api-4.0-RC03.jar:?]
at net.corda.nodeapi.internal.persistence.CordaPersistence.transaction(CordaPersistence.kt:212) ~[corda-node-api-4.0-RC03.jar:?]
at net.corda.nodeapi.internal.persistence.CordaPersistence.transaction(CordaPersistence.kt:190) ~[corda-node-api-4.0-RC03.jar:?]
at net.corda.nodeapi.internal.persistence.CordaPersistence.transaction(CordaPersistence.kt:196) ~[corda-node-api-4.0-RC03.jar:?]
at net.corda.node.internal.AbstractNode.start(AbstractNode.kt:369) ~[corda-node-4.0-RC03.jar:?]
at net.corda.node.internal.Node.start(Node.kt:419) ~[corda-node-4.0-RC03.jar:?]
at net.corda.node.internal.NodeStartup.startNode(NodeStartup.kt:180) ~[corda-node-4.0-RC03.jar:?]
at net.corda.node.internal.NodeStartupCli$runProgram$2.run(NodeStartup.kt:109) ~[corda-node-4.0-RC03.jar:?]
at net.corda.node.internal.NodeStartup$initialiseAndRun$5.invoke(NodeStartup.kt:161) ~[corda-node-4.0-RC03.jar:?]
at net.corda.node.internal.NodeStartup$initialiseAndRun$5.invoke(NodeStartup.kt:116) ~[corda-node-4.0-RC03.jar:?]
at net.corda.node.internal.NodeStartupLogging$DefaultImpls.attempt(NodeStartup.kt:445) ~[corda-node-4.0-RC03.jar:?]
at net.corda.node.internal.NodeStartup.attempt(NodeStartup.kt:116) ~[corda-node-4.0-RC03.jar:?]
at net.corda.node.internal.NodeStartup.initialiseAndRun(NodeStartup.kt:159) ~[corda-node-4.0-RC03.jar:?]
at net.corda.node.internal.NodeStartupCli.runProgram(NodeStartup.kt:107) ~[corda-node-4.0-RC03.jar:?]
at net.corda.cliutils.CordaCliWrapper.call(CordaCliWrapper.kt:184) ~[corda-tools-cliutils-4.0-RC03.jar:?]
at net.corda.cliutils.CordaCliWrapper.call(CordaCliWrapper.kt:152) ~[corda-tools-cliutils-4.0-RC03.jar:?]
at picocli.CommandLine.execute(CommandLine.java:1056) ~[picocli-3.8.0.jar:3.8.0]
at picocli.CommandLine.access$900(CommandLine.java:142) ~[picocli-3.8.0.jar:3.8.0]
at picocli.CommandLine$RunLast.handle(CommandLine.java:1246) ~[picocli-3.8.0.jar:3.8.0]
at picocli.CommandLine$RunLast.handle(CommandLine.java:1214) ~[picocli-3.8.0.jar:3.8.0]
at picocli.CommandLine$AbstractParseResultHandler.handleParseResult(CommandLine.java:1122) ~[picocli-3.8.0.jar:3.8.0]
at picocli.CommandLine.parseWithHandlers(CommandLine.java:1405) ~[picocli-3.8.0.jar:3.8.0]
at net.corda.cliutils.CordaCliWrapperKt.start(CordaCliWrapper.kt:72) ~[corda-tools-cliutils-4.0-RC03.jar:?]
at net.corda.node.Corda.main(Corda.kt:13) ~[corda-node-4.0-RC03.jar:?]
[INFO ] 2019-02-12T11:02:31,343Z [main] internal.Node.installCordaService - Installed com.r3.corda.finance.ripple.services.XRPService Corda service
[INFO ] 2019-02-12T11:02:31,390Z [main] vault.NodeVaultService.oldStatesPresent - Checking for states in vault from a previous version
[INFO ] 2019-02-12T11:02:31,406Z [main] vault.NodeVaultService.oldStatesPresent - Finished checking for old states. Old states present: false
[INFO ] 2019-02-12T11:02:34,076Z [main] statemachine.SingleThreadedStateMachineManager.invoke - Node ready, info: NodeInfo(addresses=[localhost:10011], legalIdentitiesAndCerts=[O=Oracle, L=New York, C=US], platformVersion=4, serial=1549969092158)
[INFO ] 2019-02-12T11:02:34,716Z [main] BasicInfo.printBasicNodeInfo - Loaded 6 CorDapp(s) : Combined CorDapp: Contract CorDapp: Corda Finance Demo version 1 by vendor R3 with licence Open Source (Apache 2), Workflow CorDapp: Corda Finance Demo version 1 by vendor R3 with licence Open Source (Apache 2), Combined CorDapp: Contract CorDapp: my contract name version 1 by vendor my company with licence my licence, Workflow CorDapp: my workflow name version 1 by vendor my company with licence my licence, Combined CorDapp: Contract CorDapp: my contract name version 1 by vendor my company with licence my licence, Workflow CorDapp: my workflow name version 1 by vendor my company with licence my licence, CorDapp cordapp-contracts-states-0.1 version Unknown by Unknown with licence Unknown, CorDapp oracle-0.1 version Unknown by Unknown with licence Unknown, CorDapp ripple-0.1 version Unknown by Unknown with licence Unknown
[WARN ] 2019-02-12T11:02:34,716Z [main] internal.Node.logLoadedCorDapps - Found 3 CorDapp(s) with unknown information. They will be unable to run on Corda in the future.
[INFO ] 2019-02-12T11:02:34,747Z [Node thread-1] internal.Node.registerJmxReporter - Registering JMX reporter:
[INFO ] 2019-02-12T11:02:34,763Z [Node thread-1] internal.Node.registerJolokiaReporter - Registering Jolokia JMX reporter:
[INFO ] 2019-02-12T11:02:34,794Z [main] BasicInfo.printBasicNodeInfo - Node for "Oracle" started up and registered in 151.81 sec
[INFO ] 2019-02-12T11:02:34,794Z [main] rpc.RPCServer.start - Starting RPC server with configuration RPCServerConfiguration(rpcThreadPoolSize=4, reapInterval=PT1S, deduplicationCacheExpiry=PT24H)
I neither touched the build file for the oracle nor the respective classes or the xrp.conf.
I appreciate any help! Many thanks in advance!
Here is my solution: In com.r3.corda.finance.obligation.oracle.services.XrpOracleService a small change of the code was necessary, the addition of the ClassLoader did the magic
private val nodes by lazy { ConfigFactory.parseResources(javaClass.getClassLoader(), configFileName).getStringList("nodes").mapNotNull(::URI) }
+ //private val nodes by lazy { ConfigFactory.parseResources(configFileName).getStringList("nodes").mapNotNull(::URI) }

Connect sparklyr 0.8.4 to remote spark 2.2.1 connection

I'm trying to connect from R to a remote spark cluster.
The spark cluster is build on debian jessie and the R version i can install on it is at most 3.3 but I need 3.4 to be able to run FactoMineR. So I installed R on another machine and try to connect the cluster using sparklyr 0.8.4
> sc <- spark_connect(master = "spark://spark-cluster-m:7077", spark_home="/usr/lib/spark/", version="2.2.1")
Error in start_shell(master = master, spark_home = spark_home, spark_version = version, :
SPARK_HOME directory '/usr/lib/spark/' not found
spark isn't installed on the local machine but on the spark-cluster-m, it is :
jc#spark-cluster-m:/usr/lib/spark$ ls
bin conf data examples external jars LICENSE licenses NOTICE python R README.md RELEASE sbin work yarn
Have I missed something ?
The spark cluster is on google cloud (test account) and so is the VM with R. How do I verify the port spark can be connected to ?
Thanks for your clues
#user16... You're right, this particular problem seems to be solved but my way is not ended.
I installed the same spark version (2.2.1 with hadoop > 2.7)
Here is my new error message :
Error in force(code) :
Failed during initialize_connection: java.lang.IllegalArgumentException: requirement failed: Can only call getServletHandlers on a running MetricsSystem
at scala.Predef$.require(Predef.scala:224)
at org.apache.spark.metrics.MetricsSystem.getServletHandlers(MetricsSystem.scala:91)
at org.apache.spark.SparkContext.<init>(SparkContext.scala:524)
at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2516)
at org.apache.spark.SparkContext.getOrCreate(SparkContext.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at sparklyr.Invoke.invoke(invoke.scala:137)
at sparklyr.StreamHandler.handleMethodCall(stream.scala:123)
at sparklyr.StreamHandler.read(stream.scala:66)
at sparklyr.BackendHandler.channelRead0(handler.scala:51)
at sparklyr.BackendHandler.channelRead0(handler.scala:4)
at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293)
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:267)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1294)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:911)
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131)
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:643)
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:566)
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:480)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442)
at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131)
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:144)
at java.lang.Thread.run(Thread.java:748)
Log: /tmp/RtmpTUh0z6/file5d231368db0_spark.log
---- Output Log ----
at io.netty.channel.nio.NioEventLoop.processS
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:480)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442)
at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131)
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:144)
... 1 more
18/07/21 18:24:59 INFO StandaloneAppClient$ClientEndpoint: Connecting to master spark://spark-cluster-m:7077...
18/07/21 18:24:59 WARN StandaloneAppClient$ClientEndpoint: Failed to connect to master spark-cluster-m:7077
org.apache.spark.SparkException: Exception thrown in awaitResult:
at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:205)
at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)
at org.apache.spark.rpc.RpcEnv.setupEndpointRefByURI(RpcEnv.scala:100)
at org.apache.spark.rpc.RpcEnv.setupEndpointRef(RpcEnv.scala:108)
at org.apache.spark.deploy.client.StandaloneAppClient$ClientEndpoint$$anonfun$tryRegisterAllMasters$1$$anon$1.run(StandaloneAppClient.scala:106)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.io.IOException: Failed to connect to spark-cluster-m/10.142.0.3:7077
at org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:232)
at org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:182)
at org.apache.spark.rpc.netty.NettyRpcEnv.createClient(NettyRpcEnv.scala:197)
at org.apache.spark.rpc.netty.Outbox$$anon$1.call(Outbox.scala:194)
at org.apache.spark.rpc.netty.Outbox$$anon$1.call(Outbox.scala:190)
... 4 more
Caused by: io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection refused: spark-cluster-m/10.142.0.3:7077
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(NioSocketChannel.java:257)
at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:291)
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:631)
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:566)
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:480)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442)
at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131)
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:144)
... 1 more
18/07/21 18:25:19 ERROR StandaloneSchedulerBackend: Application has been killed. Reason: All masters are unresponsive! Giving up.
18/07/21 18:25:19 WARN StandaloneSchedulerBackend: Application ID is not initialized yet.
18/07/21 18:25:19 INFO Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 46811.
18/07/21 18:25:19 INFO NettyBlockTransferService: Server created on 10.142.0.5:46811
18/07/21 18:25:19 INFO BlockManager: Using org.apache.spark.storage.RandomBlockReplicationPolicy for block replication policy
18/07/21 18:25:19 INFO BlockManagerMaster: Registering BlockManager BlockManagerId(driver, 10.142.0.5, 46811, None)
18/07/21 18:25:19 INFO BlockManagerMasterEndpoint: Registering block manager 10.142.0.5:46811 with 366.3 MB RAM, BlockManagerId(driver, 10.142.0.5, 46811, None)
18/07/21 18:25:19 INFO BlockManagerMaster: Registered BlockManager BlockManagerId(driver, 10.142.0.5, 46811, None)
18/07/21 18:25:19 INFO BlockManager: Initialized BlockManager: BlockManagerId(driver, 10.142.0.5, 46811, None)
18/07/21 18:25:19 INFO SparkUI: Stopped Spark web UI at http://10.142.0.5:4040
18/07/21 18:25:19 INFO StandaloneSchedulerBackend: Shutting down all executors
18/07/21 18:25:19 INFO CoarseGrainedSchedulerBackend$DriverEndpoint: Asking each executor to shut down
18/07/21 18:25:19 WARN StandaloneAppClient$ClientEndpoint: Drop Unregist
I can see it can resolve the name (=> 10.142.0.3)
Also, it seems to be the good port as if I use port 7000, i have the error :
18/07/21 18:32:54 ERROR TransportResponseHandler: Still have 1 requests outstanding when connection from spark-cluster-m/10.142.0.3:7000 is closed
18/07/21 18:32:54 WARN StandaloneAppClient$ClientEndpoint: Could not connect to spark-cluster-m:7000: java.io.IOException: Connection reset by peer
18/07/21 18:32:54 WARN StandaloneAppClient$ClientEndpoint: Failed to connect to master spark-cluster-m:7000
But I can't figure out what this means.
You say my configuration is "particular". If there is a better (and simple) approach, i would be glad to use it.
Here is how I proceeded in my tests :
I created a google dataproc cluster with spark (2.2.1)
I added Cassandra on each node
At this stage, everything works ok.
Then, i need to install FactoMineR as I'd like to try HMFA. It is said to run with R > 3.0.0 so it seems to be ok but it depends on nlme which can't be installed on R < 3.4.0 (and the one in the debian jessie backports is 3.3.)
So, what can I do ?
I must admit that i'm not very enthusiastic in restarting a full spark / cassandra cluster install from scratch...

bootstrap admin credentials do not exist in the config store

I am doing Aritfactory HA Installation and Setup. The primary server starts successfully, but the Secondary server always fails to start. The error is as follows:
2017-12-28 08:31:17,209 [art-init] [INFO ] (o.a.s.ArtifactoryApplicationContext:231) - Artifactory context starting up 49 Spring Beans...
Thu Dec 28 08:31:17 UTC 2017 WARN: Establishing SSL connection without server's identity verification is not recommended. According to MySQL 5.5.45+, 5.6.26+ and 5.7.6+ requirements SSL connection must be established by default if explicit option isn't set. For compliance with existing applications not using SSL the verifyServerCertificate property is set to 'false'. You need either to explicitly disable SSL by setting useSSL=false, or set useSSL=true and provide truststore for server certificate verification.
2017-12-28 08:31:17,443 [art-init] [ERROR] (o.a.w.s.ArtifactoryContextConfigListener:99) - Application could not be initialized: bootstrap admin credentials do not exist in the config store
java.lang.reflect.InvocationTargetException: null
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[na:1.8.0_151]
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) ~[na:1.8.0_151]
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[na:1.8.0_151]
at java.lang.reflect.Constructor.newInstance(Constructor.java:423) ~[na:1.8.0_151]
at org.artifactory.webapp.servlet.ArtifactoryContextConfigListener.configure(ArtifactoryContextConfigListener.java:224) ~[artifactory-web-application-5.7.1.jar:na]
at org.artifactory.webapp.servlet.ArtifactoryContextConfigListener.access$2(ArtifactoryContextConfigListener.java:186) ~[artifactory-web-application-5.7.1.jar:na]
at org.artifactory.webapp.servlet.ArtifactoryContextConfigListener$1.run(ArtifactoryContextConfigListener.java:95) ~[artifactory-web-application-5.7.1.jar:na]
Caused by: org.springframework.beans.factory.BeanInitializationException: Failed to initialize bean 'org.artifactory.security.access.AccessService'.; nested exception is java.util.NoSuchElementException: bootstrap admin credentials do not exist in the config store
at org.artifactory.spring.ArtifactoryApplicationContext.refresh(ArtifactoryApplicationContext.java:241) ~[artifactory-core-5.7.1.jar:na]
at org.artifactory.spring.ArtifactoryApplicationContext.<init>(ArtifactoryApplicationContext.java:124) ~[artifactory-core-5.7.1.jar:na]
... 7 common frames omitted
Caused by: java.util.NoSuchElementException: bootstrap admin credentials do not exist in the config store
at org.artifactory.security.access.ArtifactoryAccessClientConfigStore.lambda$14(ArtifactoryAccessClientConfigStore.java:343) ~[artifactory-core-5.7.1.jar:na]
Is there any suggestion about this error?
Thanks,
Jiang Chuan.
Fixing owner and permission of db.properties will solve your issue.

Alfresco tomcat hangs during startup

I have more or less the same problem as this person. With the exception that open/libreoffice is disabled. I have tried the bin install as well as a fresh tomcat install. All exhibit the same issue. I already had the repo instance running before hand but decided to re-install afresh as the wcmqs addon wouldn't work, since then I haven't been able to get an instance going. What else I can add is that I did notice that tomcat hangs with the CPU at 100% when I do a top inspection. There're no errors in catalina.out and this is the end of that log:
2013-05-05 01:27:28,348 INFO [management.subsystems.ChildApplicationContextFactory] [localhost-startStop-1] Starting 'imap' subsystem, ID: [imap, default]
2013-05-05 01:27:28,532 INFO [management.subsystems.ChildApplicationContextFactory] [localhost-startStop-1] Startup of 'imap' subsystem, ID: [imap, default] complete
2013-05-05 01:27:28,532 INFO [management.subsystems.ChildApplicationContextFactory] [localhost-startStop-1] Starting 'email' subsystem, ID: [email, outbound]
2013-05-05 01:27:28,557 INFO [management.subsystems.ChildApplicationContextFactory] [localhost-startStop-1] Startup of 'email' subsystem, ID: [email, outbound] complete
2013-05-05 01:27:28,557 INFO [management.subsystems.ChildApplicationContextFactory] [localhost-startStop-1] Starting 'email' subsystem, ID: [email, inbound]
2013-05-05 01:27:28,612 INFO [management.subsystems.ChildApplicationContextFactory] [localhost-startStop-1] Startup of 'email' subsystem, ID: [email, inbound] complete
2013-05-05 01:27:28,612 INFO [management.subsystems.ChildApplicationContextFactory] [localhost-startStop-1] Starting 'googledocs' subsystem, ID: [googledocs, default]
2013-05-05 01:27:28,641 INFO [management.subsystems.ChildApplicationContextFactory] [localhost-startStop-1] Startup of 'googledocs' subsystem, ID: [googledocs, default] complete
2013-05-05 01:27:28,641 INFO [management.subsystems.ChildApplicationContextFactory] [localhost-startStop-1] Starting 'Subscriptions' subsystem, ID: [Subscriptions, default]
2013-05-05 01:27:28,648 INFO [management.subsystems.ChildApplicationContextFactory] [localhost-startStop-1] Startup of 'Subscriptions' subsystem, ID: [Subscriptions, default] complete
2013-05-05 01:27:28,648 INFO [repo.usage.UserUsageTrackingComponent] [localhost-startStop-1] Disabled - clear non-missing user usages ...
2013-05-05 01:27:28,651 INFO [repo.usage.UserUsageTrackingComponent] [localhost-startStop-1] Found 0 users to clear
2013-05-05 01:27:28,652 INFO [repo.usage.UserUsageTrackingComponent] [localhost-startStop-1] ... cleared non-missing usages for 0 users
2013-05-05 01:27:28,652 INFO [management.subsystems.ChildApplicationContextFactory] [localhost-startStop-1] Starting 'Synchronization' subsystem, ID: [Synchronization, default]
2013-05-05 01:27:28,780 INFO [management.subsystems.ChildApplicationContextFactory] [localhost-startStop-1] Startup of 'Synchronization' subsystem, ID: [Synchronization, default] complete
2013-05-05 01:27:28,823 INFO [service.descriptor.DescriptorService] [localhost-startStop-1] Alfresco JVM - v1.7.0_21-b11; maximum heap size 1460,563MB
2013-05-05 01:27:28,824 INFO [service.descriptor.DescriptorService] [localhost-startStop-1] Alfresco started (Community). Current version: 4.2.0 (4576) schema 6 022. Originally installed version: 4.2.0 (4576) schema 6 022.
2013-05-05 01:27:28,824 INFO [management.subsystems.ChildApplicationContextFactory] [localhost-startStop-1] Starting 'ActivitiesFeed' subsystem, ID: [ActivitiesFeed, default]
2013-05-05 01:27:28,870 INFO [management.subsystems.ChildApplicationContextFactory] [localhost-startStop-1] Startup of 'ActivitiesFeed' subsystem, ID: [ActivitiesFeed, default] complete
2013-05-05 01:27:28,870 INFO [management.subsystems.ChildApplicationContextFactory] [localhost-startStop-1] Starting 'Replication' subsystem, ID: [Replication, default]
2013-05-05 01:27:28,877 INFO [management.subsystems.ChildApplicationContextFactory] [localhost-startStop-1] Startup of 'Replication' subsystem, ID: [Replication, default] complete
2013-05-05 01:27:30,070 DEBUG [alfresco.repo.workflow] [localhost-startStop-1] Attached JBPM Context to transaction dba830e8-cc35-4845-bd19-d4816a1a9790
2013-05-05 01:27:30,085 DEBUG [alfresco.repo.workflow] [localhost-startStop-1] Workflow deployer: Definition 'alfresco/workflow/publish_web_content_processdefinition.xml' already deployed
2013-05-05 01:27:30,088 DEBUG [alfresco.repo.workflow] [localhost-startStop-1] Workflow deployer: Definition 'alfresco/workflow/publish-web-content.bpmn20.xml' already deployed
2013-05-05 01:27:30,088 DEBUG [alfresco.repo.workflow] [localhost-startStop-1] Detached (commit) JBPM Context from transaction dba830e8-cc35-4845-bd19-d4816a1a9790
2013-05-05 01:28:00,101 WARN [alfresco.util.OpenOfficeConnectionTester] [DefaultScheduler_Worker-8] Error trying to query Open Office version information. OpenOffice.org's ConfigurationRegistry not implemented in this version of OOo. This should not affect the operation of OOo.
2013-05-05 01:28:00,102 INFO [alfresco.util.OpenOfficeConnectionTester] [DefaultScheduler_Worker-8] The OpenOffice connection was re-established.
2013-05-05 01:30:55,814 ERROR [alfresco.util.SocketOpenOfficeConnection] [MessageDispatcher] disconnected unexpectedly
2013-05-05 01:31:00,001 ERROR [alfresco.util.OpenOfficeConnectionTester] [DefaultScheduler_Worker-7] The OpenOffice connection has been lost.
I have no idea what else I can do to troubleshoot the issue.
UPDATE: I installed a repo on my local machine (OSX 10.7.5) and the system exhibits the same symptoms, which the processor being thrashed, but after about almost 30 mins the repo tomcat eventually starts up.
It turned out that there was an issue with something in the extension path. Wiping the shared/classes/alfresco/extension folder and copying over the one from my working repo made it start. Nothing was showing up in the error or debugging information so there's no way I can really say what caused this weird issue.

Resources