After a period of time airflow exceeds the slot limit defined for a particular pool. It works fine after a reboot, looking for a permanent solution for this.
Airflow version - 1.6.2
Thanks in advance.
Pool oversubscription is supposed to be a known issue up until 1.7.1.3. I think this is going to be fixed in the next release (not entirely sure though).
Related
In our company we recently upgraded to artifactory-pro:7.38.10 from version 6. To cleanup old artifacts we are using lavatory which runs an aql-search to identify the artifacts to be removed by filtering them by the date. This worked without issues our previous installation based on artifactory 6. Now after the upgrade artifactory frequently crashes with an OutOfMemoryError and the instance seems to require either significantly more memory than before or there is a memory leak. After further investigation it turned out that when problem is caused by running the aql-search and the memory usage jumps from 4 GB to over 10 GB. That's +6 GB for something that hasn't changed.
After searching for known issue I found https://www.jfrog.com/jira/browse/RTFACT-26825 which is resolved and might solve our problem but there is no version specified containing a fix. Since there is a workaround and the issue was fixed, I expect that there must be a release.
Is there already a release containing a fix?
The JIRA that you are referring to is fixed in the Artifactory version 7.38.0. So mostly, that should not be causing an issue as you are in higher version than 7.38.0.
In order to confirm, you may try the below. Add the system property to $JFROG_HOME/artifactory/var/etc/artifactory/artifactory.system.properties file and restart Artifactory for changes to take effect:
artifactory.nuget.v2.search.page.size=1000
Alternatively, you may put all the Nuget Devexpress repositories as offline. Now, check if you are encountering the memory issue. If you are not encountering this issue, maybe there is a regression issue. But according to my assumption, your server needs more resources as there are a lot of microservices introduced in Artifactory 7, when compared to Artifactory 6.
Please check if you are satisfying the resources requirement as mentioned in this page. In that case, you would need to tune your Artifactory as per this article.
I am having an issue with a relatively small openstack cluster deployed with kolla-ansible. The issue is that after a few days the controllers stop working. When I go into the docker container logs, I see in all of them that there are Too Many Open Files. I have tried changing limits.conf sysctl max files for processes and user. After all of that, the issue still shows up.
One interesting thing is that this was not happening until I had to reboot all of the controllers. I rebooted them because I needed to increase the amount of ram that they have after they died swapping. My first thought was that kolla-ansible is setting a configuration after running deploy, but I can't seem to find any point in the repo when kolla-ansible is changing ulimits or other.
Any theories what could cause this? Would it be related to increasing ram? Should I run reconfigure/deploy on each controller? I've tried looking in kolla-ansible's docs and forums and couldn't see where anyone else was having this issue.
Update this hasn't been fixed yet:
I submitted a bug report, https://bugs.launchpad.net/kolla-ansible/+bug/1901898
I don't know your used versions of Kolla-Ansible and your Linux, but your problem seems really related to this one:
On Ubuntu 16.04, please uninstall lxd and lxc packages. (An issue exists with cgroup mounts, mounts exponentially increasing when restarting container) (source: docs.openstack.org/kolla-ansible/4.0.0/quickstart.html)
I had this problem with the exponentially growing number of mount-pointers after the restart of my docker-containers too. My single-node test-deployment had become very slow based on this problem, but I can't remember at the moment, that I would had the same error with too many open files.
You can delete the packages with apt-get remove lxc-common lxcfs lxd lxd-client. I had done this fix together with a complete reinstallation of the kolla-ansible installation, so I don't know, if this also helps with an already existing installation. You should also use docker-ce instead of the docker from the apt-repos.
This was fixed with a workaround in bug https://bugs.launchpad.net/keystonemiddleware/+bug/1883659 problem was neutron server was keeping memcached connections open and not closing them until the memcached container reached too many files open. There is a work around mentioned in the bug link.
I've got an admin .aspx page that I need to run on an instance to iterate through a bunch of items and fix some Worflow fields that aren't set properly. It's got to get through a lot of items. I ran it on my dev machine and it worked fine. It took a really long time, but completed.
Now, I'm trying to run the same page in the client's environment and I keep getting an HttpException: Request timed out error.
I just need to run this thing once. Any thoughts on why it won't run there and how to make it so I can get it to complete?
In .NET There is a executionTimeout. This time-out applies only if the debug attribute in the compilation element is False. This can explain the different behavior on your dev vs production
msdn
An other option is the app pool recycle setting.
Try the following:
Open IIS, and Select "Advanced Settings" for your Application Pool
Set the Ping Maximum Response Time to a larger number, note its in seconds
Reset those values back to the original value, once the task is done
Hope This helps
On Tridion 2011 SP1, after I just restarted an HTTP Deployer, I get the error "Attempt to load JVM failed on native side" when I try to access HTTPUpload.aspx.
What is the issue?
I added an env variable JAVA_HOME, restarted the server, but no luck so far.
Many thanks in advance!
Nevermind... It seems that after rebooting the server AGAIN, problem was fixed.
I guess I'll never know what was it.
The story is way deeper than initially believed and it all boils down to memory allocation.
The culprit in my case was the heap size that we allocate to the Java process running underneath IIS (in JuggerNET). I have 4 CD instances (4 websites running each a CD stack) on 32bit server, with memory 4GB. The heap size was set to 1024M. Naturally there was not enough memory to allocate 4GB of heapspace.
Reducing the heapsize OR stopping a website solved the issue.
Heapsize is controlled in registry key
HKEY_LOCAL_MACHINE\SOFTWARE\Tridion\Content Delivery\General\jvmarg1
with value -Xmx1024M.
Another culprit might be mixing CD DLLs from 64bit with 32bit servers, so check and double check your DLLs!!! I know I did :) hours long...
Updated to xcode4 days ago, xcode4 is really nicer to xcode3. But I met a memory issue when using xcode 4. The total active memory kept growing when the xcode4 war running, grew from 500m to 2.4G, the process memory is around 200m. It's strange~
After I closed xcode, the total active memory didn't go down soon, it was 2.4G for about 10 minutes.
Has anyone else met this issue too? Thanks for any info!
== Updates ==
Upgrade to XCode4.0.2, still has memory issue
I have the same problem. At times Xcode 4 starts to index your project (you can see "Indexing" message in the status bar at the top of window). During that it could use up to 2.8GB (!) of memory.
As soon as it happens I stop to use my laptop and start to make tea :)
If the swap exceeds 500M I restart my computer. I have 4GB of memory installed in my macbook 5.2 and there is no way to increase it :(
I don't know exactly what that "indexing" actually means. I supposed that it is connected with Code Sense in some way. But when I tried to disable code completion (preferences -> text editing -> editing), it didn't help.
I hope Apple would fix it in the next release. If not, the only way is upgrade my computer. Or use Xcode 3.2.
I'm having this same issue. Currently I'm using following workaround:
I have Activity Monitor opened on a second screen, and whenever Xcode reaches 1GB I restart it, and it works smoothly once again.
I know it's far from perfect workaround, and I'm looking forward for a better one.
I have Xcode 4.0.1 & OSX 10.6.7
I found a solution!!!
I wanted to clean my /Library/Cache. Accidentally I deleted part of my /Library :-) so I decided to do a full system restore using OSX DVD and my current (20 minutes old) Time Machine backup. I did the restore and ... It fixed the problem!. Time Machine restore cleans all cache! (it should be enough if you only delete the content of /Library/Cache and {HomeDirectory}/Library/Cache). Good luck!