Skip to main content
SearchLoginLogin or Signup

Early update on Arcadia publishing 2.0: Scientists are in charge, speed is an issue

Since starting v2 of our publishing model to restore scientist agency, pub “quality” has been similar, and this approach is more efficient overall. The major downside is that pubs take longer. We’re pursuing solutions to this and other problems but feel we’re on the right track.
Published onDec 19, 2024
Early update on Arcadia publishing 2.0: Scientists are in charge, speed is an issue
·

Purpose

About six months ago, we switched to a new “version 2.0” or “v2” model for our open publishing approach. We’d strayed a bit too far from our vision of researchers rapidly sharing their science on their own terms, and we needed to remove some top-down control over timelines, content, and process. Instead, our dedicated publishing staff (“pub team”) offers support on an opt-in basis.

This pub presents the initial results from this stage of our experiment. We’ve eliminated a lot of bureaucracy, and our scientists are now fully in charge. The pub team has more time to write meta-reflection pubs (like this!), improve publishing infrastructure and build tools, help more deeply on pubs that need it, and better integrate publishing with hiring. The main problem so far is that we’re sharing less often. Without close oversight, not all of our scientists share promptly. We’ve also shifted a lot of responsibility to managers, who sometimes struggle to keep up with writing their individual pubs. Last, we haven’t seen a change in the utility of engaging with the broader community — our scientists don’t interact much externally on their own, and outside researchers seem to find and interact with our work a bit less often. Overall, though, we think these downsides are manageable, and we have plans to address them.

Keep in mind that we don’t have much data yet — we’ve only completed 11 pubs via our v2 approach. That said, we think it’s enough to provide a rough sense of how things are going. We hope anyone following our publishing experiment will appreciate the update and give us feedback! We’d love to know what you think and hear any ideas for overcoming the issues we raise.

Share your thoughts!

Watch a video tutorial on making a PubPub account and commenting. Feel free to add line-by-line comments anywhere within this text, provide overall feedback by commenting in the box at the bottom of the page, or post about this work on social media. Please make all feedback public so other readers can benefit from the discussion. 

Background on Arcadia’s evolving publishing model

We’ve been publishing work on this platform, independently of journals, since mid-2022. Two years into our experiment, we reflected on an internal challenge that had slowly emerged — the deterioration of our scientists’ agency in the publishing process — and decided to rectify it [1]. We refer to our original approach as “version 1.0” and our new strategy as “version 2.0” or “v2.”

We briefly summarize the key differences in our new v2 framework below, but check out our full pub on this evolution for more complete context.

In a nutshell, we realized that — though well-intentioned — our pub process had grown to be too complex and bureaucratic [1]. We required multiple meetings to plan and check in on pubs, and every draft had to go through review by the publishing team. In v2, we’ve shifted to a scientist-led system with just two requirements: byline contributor approval and availability of materials/information needed to reproduce the work. By default, our scientists now decide their own publishing formats, timelines, and engagement approaches without input from pub team. Our pub team still provides support services, but it's upon request from scientists, who choose from a menu of options.

We hoped this streamlined approach would better serve our goals of rapid, impactful scientific sharing while maintaining our commitment to openness and reproducibility. We figured the potential risks would be variation in quality, poor decision-making about commercialization by individual scientists, and lower pub output as a whole.

In this pub, we check how things are going with our v2 model. Read on to learn about the impacts we’ve seen so far!

The results

We’re about six months in, and we’ve applied our new v2 publishing process to 19 total pubs (10 released, one we’re keeping internal for now, and eight in progress). While we’ve been self-assessing progress constantly along the way, we thought now would be a good time to take a more comprehensive snapshot of where things stand so we can reflect and update our readers.

We’ve mapped out our progress in Figure 1, approximating the relative size of each change we’ve observed to provide an overall sense of how things have evolved. In the subsequent subsections, we discuss what’s going well and how we need to improve in more detail.

Diagram of how various facets of Arcadia’s publishing model have evolved where the biggest improvements are in areas that enhance the external replicability of the model, and the biggest emerging issue is the slow time to publish.
Figure 1

Summary of how things are going with our v2 publishing model compared to v1 as the baseline.

We’ve assigned rough ratings for how much various aspects of our publishing approach have changed in this new version of our model relative to our prior version. We group each readout based on how it connects to the big-picture goals of our publishing experiment — internal experience, speed, utility, and replicability. Keep in mind that these are approximations and only expressed relative to the original model.

What’s going well

The process is streamlined and scientists are in charge

We’ve implemented a suite of changes that automatically streamline our publishing process and shift agency back to our scientists.

Things are much more efficient. We’ve eliminated two hours of required meetings for each pub and three hours of check-in meetings between the pub team and scientific team leads each month. We still do occasional ad-hoc meetings about particular pubs or sets of related pubs, but the number of these hasn’t changed in our new approach. Unnecessary meetings are a huge opportunity cost for the company due to the people-time involved, so this reduction is a win for us.

Scientists now choose what help they get from the publishing team. Our standard v1 publishing process included about seven discrete tasks performed by members of the pub team for every pub (e.g., project management, editing the first draft, cleaning up figures to match our style, etc.). We’ve turned everything we used to do into an expanded menu of optional pub team “services.” Each pub has unique needs. Our scientists can now choose what they want and skip what they don’t. Thus far, each v2 pub receives around three services. One pub got seven services and another received none at all! Our scientists seem to have an intuitive understanding of what will help them most, and this can vary quite a bit from pub to pub (Table 1 shows our most-requested services).

Service

Description

Figure cleanup

Beautify figures, implementing style/branding guidelines

Thorough editing

Copyediting and in-depth feedback on content, structure, visuals, and overall effectiveness

Quick read-through for flow and clarity

Identify if any aspects are particularly confusing, suggest structural changes, provide feedback on visuals, and flag any key scientific details that seem to be missing

Original figures

Visualize methods, abstract concepts, organisms, processes, etc.

Technical consultation

Discuss what’s possible on PubPub and beyond, explore technical feasibility of creative ideas

First-time pub onboarding

Meet-and-greet with pub team, overview of publishing at Arcadia and services offered, etc.

Visual strategy session

Talk through how visuals can support text (including data visualization, illustrations, graphical abstracts, and number/order of figures)

Table 1. Publishing services that our scientists have thus far requested at least twice.
Listed by number of requests, with most at the top.

In addition to our scientists being empowered, their perceived experience with publishing is also improved. After we release each pub, we send an optional survey to the researchers who were involved. Scientists who’ve published through both our original and new models report that their latest experience was the same or better. That said, we’ve only released a few pubs exclusively through the v2 system so far, and don’t have many survey respondents (n = 6), so this is something we’ll have to monitor over time.

Pub “quality” doesn’t seem to have decreased

One concern about our new system was that with less oversight from our pub team, we’d start releasing pubs with more errors, confusing writing, or other issues. While “quality” is a difficult property to measure, we have a couple ways to track this, and haven’t noticed any major changes. We don’t have a ton of responses, but we haven’t seen a major change in how readers are responding to the form we include at the end of each pub asking about how clear, useful, and replicable it is. For example, based on the 214 responses we got for the pubs we released via our v1 process, each pub was rated "clear" about 85% of the time (as opposed to "a little confusing" or "confusing"). We have only 20 responses across our v2 pubs so far, but 83% of those responses called them “clear.” Responses to other questions are also largely similar across v1 and v2 pubs. We also calculate a measure of readability (the “Flesch reading ease,” or FRE, where higher numbers are easier to read — for reference, U.S. newspapers tend to have FREs around 45 to 50 [2]) for each pub using the textstat Python package (v0.7.4) [3]. The average FRE is about the same for pubs developed via either our original or new model (~43 for v1 pubs; ~40 in v2). We’ve seen potential accuracy issues with textstat’s implementation of FRE, especially related to scientific words that frequently appear in our pubs. These numbers should be consistent relative to each other, but may not reflect the true FRE of our pubs. We don’t see a strong correlation between these data points and whether or not someone used pub team services. Again, we don’t have a lot of data since we haven’t released a ton of new pubs yet and “quality” is tough to define, but we haven’t seen a huge qualitative difference either. We’ll keep tracking this.

How have we maintained this consistent quality? Well, our scientists are pretty good communicators already. And in cases where a concept is tricky to articulate well or a writer is less experienced, they do a great job of getting the help they need from their manager, a collaborator, pub team, or a combo. With more free time, the pub team is even more available to help with pubs that need extra attention.

We’ve also implemented writing evaluations as part of our hiring process. We try to hire applicants who are both philosophically aligned with our open approach to sharing and able to demonstrate their ability to communicate science clearly. It’s okay not to be a brilliant writer with perfect grammar, but the pub process is much easier for scientists who’ve mastered the fundamentals. We’ve also noticed that clarity of writing often correlates with clarity of thinking about complex topics, so it’s a helpful evaluation point even outside of the publishing context.

Last, we started using Grammarly company-wide. In addition to ensuring basic grammatical correctness and improving the structure/flow of sentences, the business-level plan lets us create custom style rules to prevent common errors specific to science writing and to ensure stylistic consistency. This replaced a huge proportion of the manual copyediting we used to do, and our scientists report other positive benefits, like making their writing more conversational, taking the second-guessing out of their process, and improving their writing quality.

We plan to experiment with AI-based writing solutions in the future, hoping not just to maintain quality, but to improve it further.

With less responsibility per pub, the pub team has more time to create core resources

Removing pub team tasks and meetings from our pub process means there’s less of a built-in burden on our team’s time. There are some clear, immediate benefits to our scientists — for example, the wait time for services is shorter, so the pub team bottlenecks pubs less often. We also have more time to invest in pubs that need close attention and help.

Having time freed up has let us reallocate it to bigger-picture pursuits that enhance our overall publishing experiment. Our publishing director has more time to help write “meta” pubs reflecting on our progress as a company (like this one!) and better incorporate writing evaluation into our hiring process. Our project manager has more time to quantitatively track the success of our experiment and build internal publishing infrastructure for our scientists (e.g., pub stats dashboards, automatic social media amplification of pubs, tools for finding relevant research, etc.). We’ve created guides on writing, accessibility, choosing the right pub type, checking first drafts for common errors, and more. We’ve also had time to implement new tools like Grammarly with extensive customized style rules and to collaboratively develop software packages to render figures in our house style [4][5].

This makes our publishing model increasingly replicable. Now that it doesn’t require a full support team, it’s much easier to imagine a lab or a small startup trying our approach. Once we migrate to the upcoming new version of PubPub, called “PubPub Platform,” we plan to share templates and many of the resources we’ve recently created to help others adapt our model.

Decision-making around commercialization has matured and hasn’t delayed sharing

We previously had an “IP check” as part of our publishing process, and this arrangement occasionally caused delays in sharing content. While legal review of pub content only took a few days at most, punting decision-making about commercialization and patentability to the end of the scientific process could occasionally cause longer delays of several weeks because essential conversations were happening too late.

In our v2 model, we’ve removed the legal check and let our scientists decide what they share and when they share it. They’re free to release pubs whenever they’re ready, but also carry more individual responsibility to think critically about how we’ll ultimately apply our research in the real world. This model helps encourage more forethought about downstream commercialization potential, during even the earliest phases of a new project.

Extricating commercialization-related strategy from publishing coincides with a similar shift elsewhere in the company. In a recent internal audit, our translation team found that rushing into pilots without planning around ideal outcomes or suitability for downstream therapeutic development was a common, preventable failure mode during their initial efforts to establish Arcadia’s startup studio. Company-wide, we’re now much more aligned on having scientists understand what it'll take to spin out a company and frontloading rigorous commercial research into their project plans.

Hearteningly, commercialization discussions haven’t delayed the release of any v2 pubs. And it isn’t just because we’re sharing all of our work without thinking it through — we’ve completed a few pubs that our scientists expediently chose to keep internal until they finish the experiments necessary to understand whether their ideas have therapeutic potential. Our v2 shift and improved recordkeeping also led us to revisit a pub we decided to hold back during the v1 period, and we realized that we should make it public. So overall, the adjustments in our practices around commercialization seem to result in a default of openness, greater efficiency in sharing, and more thoughtful discussions on the key results we shouldn’t share (yet).

One potential downside of this approach is that we’ll accidentally release information that prevents us from patenting or commercializing a discovery. This doesn’t seem to have happened but will likely only become apparent in hindsight. We’ll continue looking out for such occurrences.

Challenges

We have many ideas to address these problems, but it would extend the pub quite a bit to articulate them all. We’d love to hear any suggestions from you!

Pubs take longer, so we share less often

Let’s address the biggest issue with our new system first. It’s critical to our experiment that we share efficiently and openly at a pace that ensures our findings can benefit others quickly and that any feedback or ideas we receive are still actionable. The most worrisome trend in our v2 pubs is that they’re taking significantly longer to complete. We previously took about 59 workdays to craft a pub from start to finish (mean, n = 55, standard deviation = 33). Pubs developed through our v2 process have taken 71 workdays (mean, n = 11, SD = 37). We also have eight draft pubs at the moment, and as of this pub’s release, they’ve already been in progress for a mean time of 67 workdays. This is far too slow (for us, see below).

How long should a pub take?

It’s important to note that our pubs take much less time to write than traditional research papers. It’s hard to find reliable stats and this varies dramatically from author to author, but it can take anywhere from a few weeks to many months to write a typical manuscript [6], let alone publish it. 71 workdays from typing the first word to seeing it live online is dramatically different.

For us, though, when a pub takes a “long time” to complete, it often means the point person isn’t breaking their research down into modular “chunks” that they can share quickly, or it could indicate a scientific/performance issue. Delays also have major side effects: the process becomes inefficient with fits and starts that slow down other work, the contributors carry the nagging emotional burden of an unfinished task hanging over their heads, the external community misses out on timely information, and we can’t make use of feedback if we’ve already moved onto other projects by the time we release and get critical comments.

We’ve seen how quickly a pub can come together when there’s an external pressure, all without anyone working extra hours or feeling undue stress. So we don’t want to let ourselves slide into a norm where pubs are unnecessarily slow. Based on our experiences so far, we’d say 20–30 workdays is fast for a data-centric pub, and 35–50 is a reasonable goal. Anything beyond 50 workdays is likely counterproductive.

Why are pubs taking so long? It’s probably because the publishing team no longer provides automatic project management. Our team used to sit down with scientists and map out a plan and timeline for each pub. If someone fell behind, we’d remind them and guide them back on track. We’d also check in monthly with team leads to get a sense of all upcoming potential pubs and make sure they were kicked off in a timely manner. Without this close oversight, it’s much easier for scientists to procrastinate, and the longer a pub starts to take, the tougher it usually is to finish.

We may actually be underestimating how long pubs are taking, too. For the pub team, the most challenging part of our v2 model is a lack of visibility. We can’t tell when scientists have started writing a new pub unless they tell us or fill out our “pub kickoff” form. This happens most — but not all — of the time without prompting. We have almost no way of telling when new research could be shared, but the lead scientist, for whatever reason, isn’t writing it up. This limits our ability to track the success of our publishing experiment as a whole and makes intervention more difficult.

Many of the mechanisms we’ve developed for accountability depend on reliable tracking. For example, we have a “slow pub alert” set up for our CSO, a dashboard on a TV monitor in our headquarters that lists how long each pub has been in progress, etc. And though we no longer do automatic project management, we still provide occasional reminders. That said, none of these strategies are useful when we don’t know someone’s writing or should be writing a pub.

A few of our slowest pubs have lingered since before we switched to our v2 system. We count them as “v2 pubs” because they’ve been going through the new process, but in theory, these could have been written much earlier. Thus, some apparent overall slowness may stem from legacy pubs that suffer from problems independent of our publishing model. We hope that new pubs will move more expediently once we clear the backlog, but there are still changes we should make to speed things up.

Another factor contributing to this slow-down is that managers bear more responsibility in our v2 model, which can cause work overload.

Some managers are more overwhelmed

In our prior system, the pub team typically edited the first draft of each pub and its figures before other contributors reviewed the product. This could be slow, especially for less-polished pubs, because the pub team would spend a great deal of time puzzling through confusing information or seeking better ways to visualize data. In our new system, managers work with their reports to get first drafts into good shape before involving the pub team. Qualitatively, this appears to make the overall pub process more efficient because managers are most informed about what’s not accurate or not well-described in early pub drafts and have the domain expertise to know immediately what to suggest instead.

The flip side of this efficiency gain is that managers have much more pub work. Strategizing about and reviewing pub drafts can create hours of work each week, especially for managers with larger teams or more junior team members. We also expect managers to execute and publish research independently, so they have to keep up with both their own pubs and those of their reports. Further, we’ve had a few instances in which managers must complete a pub for someone who’s left the company, which requires scouring over old notes and code of varying quality, completing analyses, and sometimes even redoing work.

Perhaps unsurprisingly, we’ve noticed that most pubs that aren’t kicked off in a timely manner or that linger for months are those for which a manager is the point person. We don’t yet have data from reports about whether waiting on their manager to help with their pubs tends to bottleneck their progress, but we'll track this moving forward.

That said, in a quick poll of affected managers, most find it no more difficult to manage their pubs than before. All rated v2 “better” or “much better” than v1 at both a conceptual level and in practice. They also report an overall better publishing experience in the new model, consistent with company-wide data.

The usefulness of external engagement is unchanged

A final major difference between our v1 and v2 publishing models is our decision to put our scientists in charge of engaging with other researchers about their work. Why engage? We want: 1) other scientists to know about our research so they can learn and build on it; 2) to get feedback from readers who catch critical errors or provide ideas, improving or speeding up our science; and 3) for readers to benefit from insightful comments from other readers. Rather than relying on pre-publication peer review, we’ve proposed that tracking how other scientists interact with our research will reliably indicate its usefulness and validity. To this end, the pub team previously coordinated engagement efforts for each pub, and at times, we tried having a team member dedicated entirely to facilitating interactions with the scientific community. In the v2 model, we’ve put this task entirely on our scientists. We hypothesized that we’d get more useful engagement with outside experts if we made it less like “homework,” and our scientists felt the full responsibility and power of owning the process.

We’ve found that the impacts of our interactions with outside scientists are, in fact, unchanged. We send a handful of scientists to conferences every year, and our company recently held an “open house” event where we spoke more generally about our work with guests. Beyond that, scientist-driven engagement in our v2 system appears to be minimal.

While not the most reliable ways to understand how others benefit from our work, we’ve seen early signs that our v2 pubs receive fewer views on average and less interaction (visits, clones, issues) with their associated GitHub repos. However, comparisons are difficult given the small number of pubs and GitHub repositories we've released in v2. We'll need to monitor this as we release more pubs.

Tracking public comments on our pubs also suggests infrequent or less effective outreach that doesn’t lead to public feedback. We do receive new comments much more frequently than before, but it’s because in August 2024, we started requiring anyone applying to a scientific role to leave a comment as part of their evaluation. Before this, the median number of comments we received each month was four — now it’s 25! Almost all pub comments (~90%) now come from job applicants rather than readers who discovered our work organically or were made aware of it via outreach from our scientists.

Donut charts showing the percentages of different impacts assigned to public comments showing an early indication that the impact of public comments is similar for v1 pubs and v2 pubs
Figure 2

Comment impact breakdown by publishing version.

When our scientists assign impact categories to public comments, they can choose multiple categories per comment if needed. Here, we split comments tagged with multiple categories, meaning we count each individual “impact” and calculate percentages based on the total number of impacts rather than total number of comments. “Commenter got an idea” denotes a perceived benefit to the commenter themself rather than to our scientists. “Other” is any impact that appeared fewer than 10 times across v1 and v2 categorizations.

The similarity between these distributions is an early suggestion that our shift to scientist-led engagement hasn't meaningfully changed how public feedback shapes our work.

Are the comments more or less helpful now? We have scientists rate the impact of comments they receive on their research (choosing from one or more options like “No impact,” “Changed our direction/idea we’re following up on,” and “Caught typo/small error”). Though we have fewer rated comments for our handful of v2 pubs so far, the resulting impact profiles are roughly consistent between v1 (n = 223 rated comments, 231 comment impacts) and v2 pubs (n = 28 rated comments, 29 comment impacts) (Figure 2).

In summary, with less outreach to external scientists, fewer people see and build on our work. We do get more comments because we require this of job applicants, but they’re not incredibly useful outside of the hiring process.

Methods

The data we analyzed and the code we used to do so are available on GitHub (DOI: 10.5281/zenodo.14523020).

We calculate workdays in progress by recording the “kickoff date” of the pub (when we held the kickoff meeting for v1 pubs and when the kickoff form was submitted in v2) and the publication or internal completion date. We use the WORKDAY_DIFF() function in Airtable to calculate an approximate number of workdays in progress. This doesn't account for individual PTO, company events, or other external factors beyond weekends. We used this script to calculate summary statistics related to pub duration.

To generate the donut charts and associated data in Figure 2, we used data stored in Airtable and this script (modified here to accept a CSV instead of using the Airtable API) to calculate the number of times our scientists assigned a given impact. We also used the Airtable summary bar to calculate how many comments were submitted by job applicants after our switch to this requirement. The pub’s point person assigned or approved all comment impacts.

We calculated a Flesch reading-ease score for each pub using textstat (v0.7.4) in this script.

We used Grammarly Business to help clarify and streamline our text. We also used Claude (3.5 Sonnet) to suggest wording ideas and then choose which small phrases or sentence structure ideas to use, to write longer stretches of text that we edited, and to help write and debug code. We also used GitHub Copilot to help clean up code.

Key takeaways

The good

  • The process is streamlined and scientists are in charge

  • Pub “quality” doesn’t seem to have decreased

  • With less responsibility per pub, the pub team has more time to create core resources

  • Decision-making around commercialization has matured and hasn’t delayed sharing

The bad

  • The usefulness of external engagement is unchanged

  • Some managers are more overwhelmed

The ugly

  • Pubs take longer, so we share less often

Are the trade-offs worth it for us?

In a word, yes. We have a lot of ideas to improve, which we’ve omitted for space and because many are pretty specific to Arcadia’s inner workings. But even if all our attempts to improve fail completely, the benefits still outweigh the downsides (Figure 1), and we’re sure this new strategy is a move in the right direction. Scientist agency and model replicability are critical to the success of our experiment, and we’d strayed too far away from those guiding pillars. We may eventually devise yet another new model for publishing, but we won’t go back to our original approach.

That said, we’re hopeful that by implementing our improvement ideas, and perhaps by gathering constructive insights from you, our readers, we’ll be able to fix some of the bigger problems and see positive changes.

How might you benefit from what we’ve learned?

We hope you’ve found this pub interesting, but ultimately want you to take something useful away from these reflections. Even if you’re not exploring new models of research sharing, our experiment has surfaced some practical insights that might inform your publishing practices. Our high-level lessons are probably most useful for group leaders or people taking the lead on prepping a manuscript.

Balance agency with accountability

While giving scientists full control over publishing can foster creativity, some structure helps maintain momentum:

  • Consider providing more formal training in writing rather than relying purely on an apprenticeship model where junior scientists learn via observation, or, when giving feedback to authors who are learning, make sure to include comments specifically on how to improve their writing

  • Light accountability or project management may be the best compromise between heavy involvement and a hands-off approach

  • Establish clear, realistic timeframes for text and figure drafts, as well as overall completion of the manuscript

  • Ask the lead author(s) what support they think they’ll need most before you start and check in routinely to see how this changes

  • Consider tracking time spent on manuscript preparation to identify and address bottlenecks, or at least reflect on this often

  • Set up lightweight systems to flag when projects are moving too slowly or signaling problems unrelated to writing

  • Start writing sooner and include input from managers/PIs earlier to prevent a heavy lift at the end

Lean on tooling and automation

By embracing the right technological solutions, you can reduce manual workload and free up you/your team's energy for higher-value work:

  • Tools like Grammarly can help a lot — it could be worth paying for Grammarly Business and creating custom style rules for your lab or group to cut down on time spent fixing basic issues

  • Automate as much as possible! Pre-schedule reminder emails, use tools like Geekbot on Slack to get weekly updates on writing progress, create templates or scripts for quick figure design, automatically re-up social media posts about papers, etc.

What’s next?

We’ll keep tracking our publishing experiment and update you when we have new findings, ideas, recommendations, or need for input. We plan to share more data on our v2 system after we’ve released more pubs since we’ve based many of this piece's conclusions on minimal or preliminary data. We also hope to share a pub about the impact of requiring all scientist job applicants to comment publicly on our work.

We’d love to hear what you think about our initial observations and ideas to improve our 2.0 publishing model. Are there other aspects of this strategy that we should measure to understand our overall success? What should we do to address the pub slow-down and lack of external engagement?


Share your thoughts!

Watch a video tutorial on making a PubPub account and commenting. Feel free to add line-by-line comments anywhere within this text, provide overall feedback by commenting in the box at the bottom of the page, or post about this work on social media. Please make all feedback public so other readers can benefit from the discussion. 


Contributors
(A–Z)
Supervision
Visualization
Validation
Conceptualization, Formal Analysis, Supervision, Visualization, Writing
Data Curation, Formal Analysis, Methodology, Visualization
Comments
5
Mac Robertson:

When it comes to increasing engagement with publications, Arcadia could consider incorporating interactive elements, such as data visualizations, live simulations, or embedded discussion forums, directly within the articles. This would allow readers not only to engage with the content but also to explore the underlying data and models in real time. Development of these features could be added as a provided service, with these user-friendly interactive visuals acting as the “gateway figure” to the particular Pub. These interactive visuals/simulations are also very useful for improving engagement in the wider research community when shared in research spaces or science communication social media.

Additionally, developing a collaborative space where scientists can leave feedback, share insights, and suggest improvements could further enhance community-driven engagement. Offering incentives, such as recognition for contributions, the opportunity to co-author follow-up publications, or even a LinkedIn style “endorsement” system could also motivate more active participation from the scientific community.

Lastly, from my personal experience, a very good way to get scientists to engage in one another’s research and let their passions speak to one another is through events that have a more casual tone than conference, such as receptions or coffee hours. Of course these are things that often exist at conferences, but being able to harness excitement and social tone of such an event without the circumstance of a conference could have a very positive effect on scientist engagement, reminiscent of a modern-day enlightenment salon. There could even be a very straightforward incentive in this case of getting a free treat in return for commenting on another scientist’s pub, because if I learned one thing from my scientific career it’s that scientists love a free coffee!

?
Abrahim El Gamal:

In my experience, some scientists have a stronger penchant for reporting than others. While it might be counterproductive to push scientists to publish simply for the sake of publishing, utilizing project management tools to monitor project status could be a useful strategy. This approach might help the publishing group identify when a scientist is at a stage where publication makes sense. At that point, the publishing team could check in with the scientist to assess progress and, if appropriate, encourage them to start considering a publication if they haven’t already.

Megan L. Hochstrasser:

Hi Abrahim! Your thinking is similar to mine, but we’ve been unable to track projects consistently. Our organizational practices have thus far been too dynamic to accommodate that structure. We do this with teams who’ve chosen to organize themselves in a manner conducive to it, but nothing is one-size-fits-all here and even what we establish with some confidence can shift quickly. I suspect that as our company grows and we’ll end up solidifying some baseline project management and reporting norms, and then the pub team can jump in to build on those.

Robert Roth:

For more context, textstat uses Pyphen for syllable counting, which appears to have some accuracy issues (as described in this GitHub issue). We’re working on a fork that implements the CMU pronouncing dictionary with a basic fallback for words that don’t appear in it (and you can find other forks that implement various workarounds to this).

This still has some errors, but seems more accurate so far for certain words, especially ones that appear frequently in our pubs. In tests on our pubs, CMU (with fallback) errors tend to overcount syllables, while Pyphen errors generally result in an undercount.

I’d be interested to hear if anyone has experience with different syllable counting methods and tools, especially for scientific texts — I’ve been looking into other tools like SyllaPy and this regex implementation, but still see a few issues. CMU isn’t perfect either. There will always be inaccuracies as syllable counting is an inexact science, but we’d like to be as accurate as possible.

We’ll be working on this to improve its accuracy for scientific work and will release anything useful that we build.

Here are some example words, showing syllable count errors with both, comparing CMU textstat and original texstat:

Word: studying, CMU with fallback: 3, Original: 2, In CMUdict: True

Word: subtypes, CMU with fallback: 3, Original: 2, In CMUdict: False

Word: diseases, CMU with fallback: 3, Original: 2, In CMUdict: True

Word: chlamydomonas, CMU with fallback: 5, Original: 2, In CMUdict: False

Word: reinhardtii, CMU with fallback: 4, Original: 2, In CMUdict: False

Word: exhibiting, CMU with fallback: 4, Original: 3, In CMUdict: True

Word: motility, CMU with fallback: 4, Original: 3, In CMUdict: True

Word: whether, CMU with fallback: 2, Original: 1, In CMUdict: True

Word: method, CMU with fallback: 2, Original: 1, In CMUdict: True

Word: methodology, CMU with fallback: 5, Original: 4, In CMUdict: True

Word: libraries, CMU with fallback: 3, Original: 2, In CMUdict: True

Word: pipetting, CMU with fallback: 3, Original: 2, In CMUdict: False

Word: pipette, CMU with fallback: 2, Original: 1, In CMUdict: False

Robert Roth:

We know that relying solely on FRE to judge quality has its limitations, especially with scientific publications. We're also exploring other tools like the Gunning Fog Index, SMOG, and others that may be better suited to the work we release. We're interested in combining these measures with qualitative feedback — through the forms on our pubs, public comments, and conversations with internal scientists — to better understand what makes our pubs high quality from a readability perspective.

I’d be curious to hear what you think about different readability measures for science publications and other ways to measure ‘quality.’

Prachee Avasthi:

We’re trying to select for scientists who can engage with our work at a deep level and successfully participate in open dialogue. We expect our scientists to do this with their own pubs, so evaluating this skill during our hiring process is important.