Insight usage feedback article

  Today, I would like to collaborate with the community about  the real usage of Insight for Jira in world instances and provide my feedback. As we know on Cloud, Insight for Jira has a limit of  400k objects for Enterprise subscription , 60k for other situations on the Cloud. So for the on-premises releases Insight for Jira has not any limitation, except  heap size requirement.   Objects in Insight JVM memory ~10.000 4Gb ~100.000 8Gb ~500.000 16Gb ~1.000.000 32Gb ~2.000.000 64Gb ~5.000.000 128Gb Note the next info: remember to always test the memory consumption in a test environment for a huge data set, because it's not always the number of objects, but the content of the object attributes as well at play.  i.e. one of my instance has  Jira/JSD/JSM version:  8.13.9 (3 nodes in DC) Insight version:  8.7.9 Java version:  11.0.11+9 GC Strategy:  G1GC   Heap Size:  38 GB Count of Objects:  1.6 mln objects Count of Units (Attribute values):  49 mln units At the moment, it works well, b

Story about on-premise installation should be reviewed on OS level?

  Hello,  Today I would like to share a small story about the performance improvement of Confluence for one large company. They used that installation for more than 3.2k requests per minute based on the daily stats from frontend reverse proxy logs. The architecture of that is quite simple based on the reverse proxy (nginx),  Confluence app (tomcat), PostgreSQL as RDMS.  All those services work on CentOS 7.  On Confluence in (Tomcat logs and application logs), nothing informative logs were.  So in dmesg I found quite simple information about SYN flood on the Confluence app side. Of course, it’s quite crazy to see that situation in an organisation in  2020 ’s. # dmesg | tail [  734.711105] systemd[1]: Started Journal Service. [1140053.637848] FS-Cache: Loaded [1140053.662442] FS-Cache: Netfs 'cifs' registered for caching [1140053.662535] Key type cifs.spnego registered [1140053.662538] Key type cifs.idmap registered [1140053.662889] Unable to determine destination address. [11400

How to set "Log work" button on the panel of issue view instead of under "More"

  Hi!  Today I would like to share just one of the customisations related to the UI. Of course, it's a small AJS, AUI trick :)   Still, most of my Jira instances are DC and Server.  And quite often request to see the standard Log work button on the next panel.  instead of here   That’s easy to set via console log: snippet: const logButton = '<a id="new-log-work" title="Log work" class="aui-button toolbar-trigger issueaction-log-work" href="/secure/CreateWorklog!default.jspa?id='+JIRA.Issue.getIssueId()+'" resolved=""><span class="trigger-label">Log Work</span></a>'; AJS.$('#opsbar-opsbar-operations').append(logButton);    We can set that change into an Announcement  banner  or use some app,  like JSincluder.  You can in marketplace or in github.   Hope it helps. Cheers, Gonchik

Insight as large Asset Management has to have retention policy

  Hello community,    Today, I would like to share a small snippet sql request to cleanup Insight object histories.  In the future, I will be happy to see the configuration in the Insight app as a retention policy. ( Please ,  Atlassian  provide retention policy feature in Insight)   We have 2mln+ objects in Insight, not so far, at the moment we do a cleanup and review existing potential weaknesses and improve performance of our Jira Service management DC setup.  Once our web navigation Insight tab was slow, If we do cleanup of history it starting to be fast then do review of records.  Every night our system syncs up with SAP, IdP and other external systems, also, event based sync up and updates objects from remote systems.  Therefore potential weakness is duplicated history, because change log we don’t need to keep longer than 3 months, as change log keeps on master systems.  Also, the next query helpful to find misconfigs.  SELECT hist."ID", hist."AFFECTED_ATTRI

My top queries to analyze the DB on PostgreSQL

Hi community,    Today I would like to share a small snippet to investigate DB and RDBMS PostgreSQL.  Before starting to check and do “explain analyze” of slowest and frequent slow queries I do: select sum(xact_commit) as commits, sum(xact_rollback) as rollbacks from pg_stat_database; Just to understand how often we  meet with rollbacks. I do recommend to run a few times that query with a small timeout to understand frequency.    Next one is used and remained connections  query : select max_conn,used,res_for_super,max_conn-used-res_for_super res_for_normal from  (select count(*) used from pg_stat_activity) t1,  (select setting::int res_for_super from pg_settings where name=$$superuser_reserved_connections$$) t2,  (select setting::int max_conn from pg_settings where name=$$max_connections$$) t3; After we are ready to check the  status of connections . Helpful to plan pooling. with states as    (select datname, client_addr, case        when now() - state_change < interval '10 seco

How only 2 parameters of PostgreSQL reduced anomaly of Jira Data Center nodes

  Hello community, Gonchik Tsymzhitov is in touch. Today, I would like to share with you a short story about in usage.  So from previous articles I hope you configured the PostgreSQL server and  After executing command, ./ --ssd for ssd disks server. You can see like this output: version 1.0.1 [OK]      I can invoke executables Connecting to /var/run/postgresql:5432 database template1 as user 'postgres'... [OK]      The user account used by me for reporting has superuser rights on this PostgreSQL instance =====  OS information  ===== [INFO]    OS: linux Version: 5.4.17-2036.102.0.2.el8uek.x86_64 Arch: x86_64-linux-thread-multi [INFO]    OS total memory: 41.03 GB [OK]      vm.overcommit_memory is adequate: no memory overcommitment [INFO]    Running under a vmware hypervisor [INFO]    Currently used I/O scheduler(s): mq-deadline =====  General instance informations  ===== -----  PostgreSQL version  -----

История о том, что старые и надежные утилиты порой нуждаются troubleshooting на новых OS (CentOS 8)

  Привет сообщество,  на связи Гончик Цымжитов. Сегодня, хочу поделиться с короткой историей, о том, как верный инструменты неожиданно ломают на новых ОС, и как легко чинятся. Итак, все мы знаем, что , через очень промежуток времени, поэтому я перехожу везде на Oracle Linux 8.  Мой любимый инструмент по настройке серверов PostgreSQL, это, где я уже рассказывал в этой статье .  Настройка как обычно идет очень просто: # dnf install -y perl-DBD-Pg # wget -O # chmod +x Но в этом, раз когда запускал скрипт встретился с такой проблемой, # ./ --ssd Can't locate in @INC (you may need to install the Memoize module) (@INC contains: /usr/local/lib64/perl5 /usr/local/share/perl5 /usr/lib64/perl5/vendor_perl /usr/share/perl5/vendor_perl /usr/lib64/perl5 /usr/share/perl5) at ./ line 262. BEGIN failed--compilation aborted at ./ line 262. Д