Evergreen ILS Website

IRC log for #evergreen, 2016-05-20

| Channels | #evergreen index | Today | | Search | Google Search | Plain-Text | summary | Join Webchat

All times shown according to the server's local time.

Time Nick Message
00:02 bshum Oh that's odd
00:03 bshum Why'd I think it was only 200000, the README clearly says 2000000 :(
00:03 * bshum is having one of those nights
00:03 bshum Nevermind then, all is right in the world, except that bshum can't read clearly anymore
00:11 bshum I decided to just do a signoff branch instead of committing, better safer when someone else double-checks it again tomorrow.  And maybe test it with Debian versions too.
00:11 * bshum calls that a good night
00:25 mceraso joined #evergreen
02:27 gmcharlt_ joined #evergreen
06:29 geoffsams joined #evergreen
07:09 rjackson_isl joined #evergreen
07:22 agoben joined #evergreen
07:38 kmlussier joined #evergreen
07:40 kmlussier rhamby++ # Mac client
07:57 krvmga joined #evergreen
08:03 kmlussier bshum: Yeah, we noticed the modperl issue last week when my VM builds starting failing.
08:08 csharp @blame mod_perl
08:08 pinesol_green csharp: It's all mod_perl's fault!
08:13 kmlussier Good morning csharp!
08:16 bshum kmlussier: Well that's special.
08:17 bshum Might have to make a separate branch for rel_2_4 OpenSRF for that then. Till we move all those bits to Evergreen side anyways per gmcharlt's wishlist to change
08:34 Dyrcona joined #evergreen
08:40 * Dyrcona starts the process of building four vms.
08:41 mmorgan joined #evergreen
08:49 jeff does it go something like this? ansible-playbook -i util/ec2.py util/provision-ec2-eg-dr.yml
08:50 Dyrcona jeff: :P
08:50 Dyrcona of course it doesn't....
08:50 Dyrcona It goes like this ./buildvm.sh
08:51 jeff automation is automation, and automation is good. :-)
08:52 Dyrcona Yep.
08:54 kmlussier @quote random
08:54 pinesol_green kmlussier: Quote #109: "< kmlussier> I live for sanity checks." (added by csharp at 10:06 AM, April 08, 2015)
08:54 Dyrcona I am reminded of an excerpt from the book Practical Common Lisp: http://www.gigamonkeys.com/book/macros-defining​-your-own.html#the-story-of-mac-a-just-so-story
08:54 Dyrcona @quote random
08:54 pinesol_green Dyrcona: Quote #18: "Alan McKay: Like the ski resort of girls looking for husbands and husbands looking for girls, the situation is not as symmetrical as it might seem." (added by Dyrcona at 11:51 AM, November 10, 2011)
08:54 Dyrcona hah. Not from the channel, but I like it. :)
08:59 mrpeters joined #evergreen
09:06 mmorgan joined #evergreen
09:17 bos20k joined #evergreen
09:20 yboston joined #evergreen
09:26 afterl joined #evergreen
09:41 Dyrcona @blame ubuntu
09:41 pinesol_green Dyrcona: ubuntu caused the white screen of death!
09:42 Dyrcona Don't think I should be seeing this while installing prerequisites: grep: /etc/apache2/httpd.conf: No such file or directory
09:42 bshum They probably did... those jerks.
09:43 csharp Dyrcona: that's from running the EG Makefile.install for xenial?
09:44 Dyrcona csharp: Nope, OpenSRF/src/extras/Makefile.install ubuntu-trusty
09:45 Dyrcona And, I only started the trusty VMs, so....
09:45 csharp hmm
09:45 tsbere I think they have been working to switch from httpd.conf to apache2.conf for a while, maybe they finally stopped putting httpd.conf in by default?
09:45 Dyrcona Maybe.
09:45 csharp if so, that's a crappy thing to do in an LTS
09:45 * Dyrcona echoes csharp
09:46 tsbere More specifically, I don't think it has been *read* by default for a while, so...
09:46 Dyrcona apache2.conf and httpd.conf are bother there.
09:46 Dyrcona s/bother/both/
09:46 Dyrcona "Oh, bother," said Pooh.
09:47 tsbere Ok, that is a very different question. Mainly "why can't grep see httpd.conf then?"
09:47 Dyrcona Maybe it wasn't there when the grep ran?
09:47 Dyrcona It was possibly installed after.
09:47 Dyrcona I am root or I would not have made it that far. :)
09:48 Dyrcona Looking at what it was doing, it may not be important.
09:48 Dyrcona I'm building on a concerto vm so I can test one of my own patches real quick.
09:49 Dyrcona I was going to test berick's angular fixes a bit later.
09:51 Dyrcona I'll probably just use concerto today, since I don't feel like waiting on reingests, etc. on a copy of production.
09:53 Dyrcona bshum: You're right about trusty not activating mod_perl any more.
09:54 mmorgan joined #evergreen
09:55 Dyrcona Well, opensrf and opensrf.math add work.
09:55 Dyrcona Now to install and test Evergreen.
10:00 berick joined #evergreen
10:04 afterl left #evergreen
10:06 bos20k_ joined #evergreen
10:06 jvwoolf joined #evergreen
10:16 * jeff shakes fist at search path
10:18 Dyrcona On the plus side, I'll be pushing a working branch for LP 1548993 shortly.
10:18 pinesol_green Launchpad bug 1548993 in Evergreen 2.9 "TPAC Show More/Fewer Details Button does not work with show_more_details.default set to true" [Undecided,Confirmed] https://launchpad.net/bugs/1548993
10:24 jeff mumble grumble functions with hardcoded references to schemas that no longer exist...
10:26 Dyrcona Or maybe not so shortly....
10:29 mmorgan1 joined #evergreen
10:30 tsbere jeff: I feel your pain. Though I wonder if these are custom functions or not.
10:47 csharp @decide I am root or I AM GROOT
10:47 pinesol_green csharp: go with I am root
10:49 tsbere @decide Blame Vendors or Blame Users
10:49 pinesol_green tsbere: go with Blame Vendors
10:51 Dyrcona Or maybe about a couple of minutes ago... ;)
11:03 jeff tsbere: non-custom afaik, but likely mangled over time by weird search path issues and numerous migrations.
11:26 bmills joined #evergreen
11:30 kmlussier @dessert
11:30 * pinesol_green grabs some Key Lime Pie for kmlussier
11:30 kmlussier pinesol_green: That's not what I had in mind.
11:30 pinesol_green kmlussier: Have you tried throwing it across the room?
11:30 pinesol_green kmlussier: I am only a bot, please don't think I'm intelligent :)
11:39 Stompro joined #evergreen
11:45 kmlussier @marc 263
11:45 pinesol_green kmlussier: The projected date of publication used in bibliographic records for works that have not yet been published. []
11:48 bmills joined #evergreen
11:49 bmills1 joined #evergreen
12:07 brahmina joined #evergreen
12:21 jihpringle joined #evergreen
12:28 mmorgan joined #evergreen
12:32 bmills joined #evergreen
12:55 mtcarlson left #evergreen
13:00 krvmga joined #evergreen
13:29 * jeff reads pingest.pl to determine if --skip-browse --skip-search --skip-facets is essentially a attrs-only reingest
13:31 jeff looks like "yes"
13:33 Dyrcona It is.
14:04 kmlussier So, for the purposes of posting the fix to the lp957466_update_date_and_source.pg test, should I re-open bug 1447746, which broke the test, or should I open a new bug?
14:04 pinesol_green Launchpad bug 1447746 in Evergreen "Do not update bib source on match-only merge" [Wishlist,Fix committed] https://launchpad.net/bugs/1447746
14:09 * kmlussier decides to re-open the old bug.
14:09 Dyrcona kmlussier: bshum has been reopening old bugs for such things. I tend to open a new bug.
14:09 Dyrcona Reopening the old bug seems fine to me.
14:10 Stompro bmills++ ; pointing out the series facet fix presented at the conference to me.
14:15 Stompro miker++ for the Metadata Abattoir presentation, thanks.
14:27 jeff Dyrcona: thanks for the confirmation. :-)
14:30 Dyrcona Well, looks my xenial vms don't really work.
14:30 Dyrcona Not sure what's causing the breakage beyond network interface names changing, but all kinds of thing seem wrong with them.
14:31 berick built my first functioning xenial image for the first time yesterday
14:32 berick only issue I had was (once again) requiring the use of apache2ctl-websockets to stop/start websockets. init/service scripts no likey
14:32 berick probably something that could be fixed w/ a small amount of effort
14:36 Dyrcona berick: how are you building your images? I'm using python vmbuilder, and I think this is a recent issue.
14:37 berick Dyrcona: virt-manager
14:38 Dyrcona berick: You're using an ISO?
14:38 * jeff gives up on a record attr ingest after only 3 hours, opts to give pingest a try
14:38 berick Dyrcona: yeah, 16.04 server ISO as of yesterday
14:39 Dyrcona jeff: pingest will still probably take longer than 3 hours.
14:40 Dyrcona berick: It was working for me when I built from an ISO with virt-install and then use virt-viewer or virt-manager to finish the installation.
14:41 Dyrcona berick: I also think it was working with python vmbuilder when I last built on in late April.
14:41 * berick nods
14:46 Dyrcona eg_db_config also hangs on me.
14:47 Dyrcona And, I'm not creating the database or loading data.
14:47 Dyrcona Just --update-config --service-all --create-offline and the options for the database user, host, port, and database.
14:48 Dyrcona Ah, crazy....
14:48 Dyrcona It works by hand, but not from a shell script...
14:48 jeff broken name resolution?
14:48 jeff hrm.
14:49 berick hm, yeah, double check /etc/hosts
14:50 berick for private.localhost / public.localhost
14:50 Dyrcona That's all set up.
14:50 Dyrcona The execscript seemed to work.
14:51 Dyrcona As I mentioned earlier, I had to change the network config to ens3 from eth0.
14:51 Dyrcona Starting services, all seems well.
14:54 Dyrcona It seems to work. I can log in with the web staff client and look up patrons.
14:55 Dyrcona And, that was all I really wanted to do.
14:55 * Dyrcona was testing berick's angular fix on trusty and on xenial.
14:56 Dyrcona I asked in #ubuntu-virt, but I don't expect any answers.
14:56 * Dyrcona always has the special problems.
15:00 * Dyrcona is going to push the fixes to master. They work AFAICT.
15:03 jeff i suppose if i'm this impatient i should be benchmarking in smaller batches than "everything"
15:03 berick Dyrcona++
15:05 pinesol_green [evergreen|Bill Erickson] LP#1554714 Ang. 2.5 hotkeys plugin update - <http://git.evergreen-ils.org/?p=​Evergreen.git;a=commit;h=3e61494>
15:05 pinesol_green [evergreen|Bill Erickson] LP#1554714 Add explicit phantomjs-prebuilt dependency - <http://git.evergreen-ils.org/?p=​Evergreen.git;a=commit;h=2f024d4>
15:09 jeff berick++ Dyrcona++
15:09 berick bshum++
15:10 jeff i (mostly idly) wonder why our config.metabib_field def for Genre (id=33) was indexing 659 in addition to 655
15:10 jeff we have two records with 659 tags.
15:11 jeff (659 is not defined)
15:14 kmlussier jeff: It seems like I asked that question at one point, either on the bug or IRC. I can't recall the answer now.
15:15 jeff oh, i thought it was purely a local quirk!
15:15 jeff i hadn't even bothered to troll through git history.
15:15 kmlussier jeff: Oh, is that a local index? Maybe I'm thinking of something else then
15:15 kmlussier jeff: https://bugs.launchpad.net/ever​green/+bug/1067823/comments/10
15:15 pinesol_green Launchpad bug 1067823 in Evergreen "tpac: genre links in record details page launch subject search" [Medium,Fix released]
15:16 jeff well, we had defined it at some point in the past, but the upgrade scripts from 2.7.4 to 2.10 define it fresh, and the only difference between that and what we already had was our xpath included 659.
15:16 * jeff looks at the bug
15:16 jeff istr we backported the index def, so we probably did that before the change noted in that bug.
15:17 kmlussier jeff: Before the index was created in master, the OPAC display of the genre fields had always included the 659, IIRC
15:17 kmlussier So if you were basing your index on what was used in the display, I can see why you would have included the 659
15:18 kmlussier Always, being as far back as we had the tpac. I don't know what we did in jspac
15:32 jvwoolf1 joined #evergreen
15:38 Christineb joined #evergreen
16:00 kmlussier @dessert
16:00 * pinesol_green grabs some Pumpkin Pie for kmlussier
16:03 * bshum likes pumpkin pie.
16:04 bshum berick++ Dyrcona++ # yay for the return of working webstaff!
16:05 Dyrcona bshum++ berick++
16:05 Dyrcona @dessert
16:05 * pinesol_green grabs some Krispy Kreme Donuts for Dyrcona
16:06 bshum Now if we can test and merge kmlussier's fix for the pgtap tests then maybe we'll get a live test success again!
16:06 bshum Weekend project, whee!
16:07 kmlussier I like pumpkin pie too, but I've mostly been looking for chocolate today.
16:08 kmlussier pinesol_green has been holding out on me.
16:08 pinesol_green kmlussier: What do you mean? An African or European swallow?
16:08 pinesol_green kmlussier: I am only a bot, please don't think I'm intelligent :)
16:12 berick @eightball is pinesol_green a bot?
16:12 pinesol_green berick: _I_ don't know.
16:12 bshum Spooky.....
16:13 berick pinesol_green just passed the Turing test
16:13 pinesol_green berick: Thank you berick! But our princess is in another castle!
16:13 pinesol_green berick: I am only a bot, please don't think I'm intelligent :)
16:17 csharp @blame [quote random]
16:17 pinesol_green csharp: Quote #57: "< jeff_> nine million useless rows in money.billing, nine million useless rows... take one down, pass it around, eight million nine-hundred and ninety-nine thousand nine-hundred and ninety nine useless rows in money.billing..." (added by csharp at 03:15 PM, May 23, 2013) wants the TRUTH?! Quote #57: "< jeff_> nine million useless rows in money.billing, nine million useless rows... (1 more message)
16:17 csharp @more
16:17 pinesol_green csharp: take one down, pass it around, eight million nine-hundred and ninety-nine thousand nine-hundred and ninety nine useless rows in money.billing..." (added by csharp at 03:15 PM, May 23, 2013) CAN'T HANDLE THE TRUTH!!
16:19 Dyrcona @quote random
16:19 pinesol_green Dyrcona: Quote #135: "<Dyrcona> ASCII stupid question; get a stupid ANSI. :)" (added by kmlussier at 09:41 AM, December 22, 2015)
16:20 kmlussier I added that quote? huh
16:23 jeff Oh hey! Those nine million useless rows are dead now. That's happy. :-)
16:25 Dyrcona jeff++
16:30 berick speaking of big deletes, I'm looking forward to deleting 260M rows from usr_activity soon and making it all transient.
16:31 berick and by deleting, i mean truncating the table, after saving a tiny fraction of the data
16:32 berick having a load balancer log in to 3 different sip machines once per second each (to confirm they are responding) really blows up the activity table, yeesh
16:32 jeff heh.
16:33 jeff i think the most frequent automated SIP login we have is every 5 mins.
16:34 jeff so, sounds like we'd have about 900 times fewer rows in said table :-)
16:35 jeff okay, 288 minutes is more than i'm willing to let this run for today. time to kill and do some more benchmarking on smaller batches.
16:36 jeff i suspect i'll benefit from tuning some postgres options AND from pingest.
16:38 jeff oh, wait.
16:38 jeff ah, okay. for a half moment i thought it had just finished. :-)
16:52 jeff yeah, tuning required.
16:53 jeff 373.918 ms vs 4704.191 ms for ... ten records.
16:53 jeff (which is almost too small a sample for reasonable benchmarking)
17:16 mmorgan left #evergreen
17:18 jeff well nuts. one of these things is not like the other.
17:22 jeff YMMV (YDBMV?), but metabib.reingest_record_attributes between 2.7 and 2.10 appears to be ~12x slower
17:24 jeff a big part of that might be the new rows in config.marc21_ff_pos_map.
17:32 jeff and config.coded_value_map
17:35 bmills joined #evergreen
17:52 dbwells jeff: Our initial ingest after the 2.9 upgrade in December was nice and fast.  Recent partial re-ingest (still on 2.9) was dramatically slower.  Haven't had time to look into it any further, but glad to think we might not be alone.
17:54 jeff dbwells: any point release upgrades between those two ingests?
17:55 jeff dbwells: and what kind of partial re-ingest were you dealing with?
17:56 dbwells jeff: I don't think so for point upgrades.  More likely a point upgrade for Postgres than for EG, if anything.
17:58 dbwells jeff: It was just a full reingest of a few hundred records, but slow enough to notice.  I've really just been ignoring it and hoping my testing was off that day for some other reason.
17:59 jeff heh
18:00 jeff it looks like with the help of pingest i can do a record-attrs-only ingest of ~400k records in 6 hours (16 pingest children)
18:00 jeff extrapolating from 2000 records in 100 record batches with 4 and 16 children.
18:00 jeff 4m2.880s with 4 children
18:01 jeff 1m48.872s with 16 children
18:02 jeff (for 2000 bibs)
18:08 dbwells jeff: At 6pm on Friday, you may also find yourself in figure-out-later land :)  Have a good weekend!
18:11 jeff dbwells: you too!
18:12 jeff dbwells: I don't know what might explain what you encountered in 2.9, but I suspect the speed difference I'm seeing is mostly due to the additions that 0967 makes to config.marc21_ff_pos_map, config.record_attr_definition, config.coded_value_map, and config.composite_attr_entry_definition
18:14 jeff as part of bug 1371647
18:14 pinesol_green Launchpad bug 1371647 in Evergreen "config.marc21_ff_pos_map needs an audit" [Wishlist,Fix released] https://launchpad.net/bugs/1371647
18:16 dbwells I am perfectly willing to believe my testing was a fluke of some kind, though not looking forward to things getting even slower, if that's the case.
18:21 jeff those four tables gained:
18:21 jeff 128, 782, 2610, and 105 rows
18:22 jeff so yeah, I think it's reasonable to expect that to impact speed of ingest.
18:56 jeff are TADL and Sitka the first to take the 2.10 plunge?
19:44 jeffdavis I should take a look at pingest.pl, I'm accustomed to doing reingests more or less manually
19:47 jeffdavis I've been seeing ~48hrs for a record attribute reingest of 1.8M records with 6 parallel processes in 2.10, but that's not really an apples-to-apples comparison with pingest.pl
19:58 jeff I'm currently doing an attribute-only reingest using pingest: pingest.pl --skip-browse --skip-search --skip-facets --max-child 28
19:59 jeff it might not be that far off from how you're doing things in parallel.
20:03 jeffdavis (1) psql -c 'SELECT id FROM biblio.record_entry WHERE NOT deleted' -o id-list.txt
20:03 jeffdavis (2) split -d -n l/6 id-list.txt reingest.sql.
20:04 jeffdavis (3) open 6 screen windows, connect to the db, and run reingest.sql.XX separately in each
20:05 jeffdavis er (2.5) `sed -i 's/^\(.*\)$/SELECT metabib.reingest_record_attributes(\1);/' reingest.sql.*`
20:06 jeffdavis so I guess effectively the same, but much more cumbersome my way :)
20:08 jeff i think pretty close, yep. :-)
20:09 jeff pingest uses a slightly different query, but any differences in speed are probably small.
20:10 jeff for the record attrs: SELECT metabib.reingest_record_attributes(id, NULL::TEXT[], marc) FROM biblio.record_entry WHERE id = ?
20:10 jeff where the ? is a placeholder that gets a list of (by default) 10,000 bre ids.
20:11 jeff oh, actually no. it runs one query per record id.
20:11 jeff i misread.
20:15 jeff jeffdavis: are you aware of any other systems that have upgraded to 2.10 yet?
20:16 jeff jeffdavis: looks like Sitka and TADL are both upgrading this weekend, and beyond that I'm not sure if anyone else has gone before or has published a schedule.
20:16 jeff (I didn't spend a LOT of time looking.)
20:18 jeffdavis1 joined #evergreen
20:18 jeffdavis1 bah, irc problems :(
20:18 jeff doh.
20:18 jeffdavis1 jeff: I'm not aware of anyone else on 2.10 yet
20:18 jeffdavis1 although I could easily have missed someone
20:18 jeff Apparently I was overly optimistic about someone else having already gone first by now when I picked this date back in March or April. :-)
20:18 jeffdavis1 fwiw things have gone really smoothly in our testing, so I'm cautiously optimistic
20:19 jeff excellent.
20:20 jeffdavis1 you're on PG 9.4 as well, right?
20:20 jeff correct.
20:22 jeff Is your upgrade window duration mostly to ensure that you have enough time for the ingest to fully complete before the system's back under production load, or are you anticipating other bits to take a lot of time also?
20:22 jeff (or rolling other things into the downtime, like hardware upgrades/etc)
20:25 jeffdavis1 it's mostly to give time for testing before we go live
20:25 jeff got it.
20:26 jeff multiple reasons for asking: curiosity / comparing notes, as we're upgrading this weekend also
20:26 jeff but also, i'm thinking about ways overall to make upgrades less painful.
20:27 jeffdavis1 I actually think a 24 hour window would have been more than enough for us this year, but it was scheduled before we had a good sense of what was going to be involved
20:27 jeffdavis1 and most of our libraries are closed on mondays anyways
20:27 jeff ah.
20:28 jeff yeah, saturday nights are good for us to start -- last libraries close at 6 PM, then we have 18 hours until the next open time (noon sunday)
20:29 jeff that's a single location, but our largest. good chance if things go wrong we'll notice.
20:29 jeff then there's another 5pm - 9am close time and we have a handful of libraries open monday, then on tuesday everyone's open.
20:31 jeff we put patron auth for most electronic resources in offline mode while we're down, maintenance pages up for the catalog (need to work on some better options for the mobile apps), and then bring things back online when we're ready.
20:32 jeffdavis1 yup, same process here
20:33 jeffdavis1 I'm hoping things go well enough that we are ready to go live on Monday with the smaller number of libraries, although part of the issue there is having support staff on hand to answer calls/issues if anyone reports issues
20:33 * jeff nods
20:33 jeffdavis1 oh, it's also a long weekend in Canada this weekend, not sure if that's the case for you folks
20:34 jeff ah, nope. that's next weekend for us.
20:35 jeff The last Monday in May is Memorial Day.
20:36 jeffdavis1 ah right
20:37 * jeff ponders crazy things like pre-computed re-ingests
20:38 jeff if you re-ingest beforehand on a snapshot of production, then dump the tables in question, you could truncate-and-copy in production after the upgrade, then reingest just those bibs that have been modified since your snapshot...
20:38 jeff you'd need to adjust sequences and such...
20:39 jeff but i don't think you'd run into major foreign key issues...
20:43 jeff probably a few options to try with the indexes on the tables.
20:44 jeff but overall i suspect that you might save a bit of time.
20:44 jeff (or, spend the time beforehand)
20:44 jeff to be a little more precise.
20:49 jeff might only be worth the effort on a sitka or pines or indiana scale.
20:54 jeffdavis1 an idea worth exploring for sure
22:40 jeff pingest winding down...
22:51 jeff oh, interesting. the default batch size was probably too large for my situation.
22:51 jeff because now it's down to 11 batches and therefore only using 11 cores.
22:52 jeff but at 10,000 records per batch, that's about... 83 minutes left.
22:54 jeff (at half power or so)
23:05 Stompro joined #evergreen
23:06 gsams joined #evergreen

| Channels | #evergreen index | Today | | Search | Google Search | Plain-Text | summary | Join Webchat